diff --git "a/ast_indexer" "b/ast_indexer" new file mode 100644--- /dev/null +++ "b/ast_indexer" @@ -0,0 +1 @@ +{"index": {"('MODELS', 'fill-mask', 'ponet')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/ponet/fill_mask.py", "imports": ["transformers", "torch"], "module": "modelscope.models.nlp.ponet.fill_mask"}, "('MODELS', 'document-segmentation', 'ponet-for-document-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/ponet/document_segmentation.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.ponet.document_segmentation"}, "('MODELS', 'extractive-summarization', 'ponet-for-document-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/ponet/document_segmentation.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.ponet.document_segmentation"}, "('MODELS', 'backbone', 'ponet')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/ponet/backbone.py", "imports": ["transformers", "math", "packaging", "torch", "distutils"], "module": "modelscope.models.nlp.ponet.backbone"}, "('HEADS', 'text-generation', 'text-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/text_generation_head.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.heads.text_generation_head"}, "('HEADS', 'token-classification', 'token-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/token_classification_head.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.heads.token_classification_head"}, "('HEADS', 'named-entity-recognition', 'token-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/token_classification_head.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.heads.token_classification_head"}, "('HEADS', 'part-of-speech', 'token-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/token_classification_head.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.heads.token_classification_head"}, "('HEADS', 'information-extraction', 'information-extraction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/infromation_extraction_head.py", "imports": ["torch"], "module": "modelscope.models.nlp.heads.infromation_extraction_head"}, "('HEADS', 'relation-extraction', 'information-extraction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/infromation_extraction_head.py", "imports": ["torch"], "module": "modelscope.models.nlp.heads.infromation_extraction_head"}, "('HEADS', 'text-classification', 'text-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/text_classification_head.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.heads.text_classification_head"}, "('HEADS', 'sentence-similarity', 'text-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/text_classification_head.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.heads.text_classification_head"}, "('HEADS', 'nli', 'text-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/text_classification_head.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.heads.text_classification_head"}, "('HEADS', 'sentiment-classification', 'text-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/text_classification_head.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.heads.text_classification_head"}, "('HEADS', 'fill-mask', 'roberta-mlm')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/torch_pretrain_head.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.heads.torch_pretrain_head"}, "('HEADS', 'text-ranking', 'text-ranking')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/text_ranking_head.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.heads.text_ranking_head"}, "('HEADS', 'fill-mask', 'bert-mlm')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/fill_mask_head.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.heads.fill_mask_head"}, "('HEADS', 'fill-mask', 'fill-mask')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/fill_mask_head.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.heads.fill_mask_head"}, "('HEADS', 'fill-mask', 'xlm-roberta-mlm')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/fill_mask_head.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.heads.fill_mask_head"}, "('HEADS', 'token-classification', 'lstm-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'named-entity-recognition', 'lstm-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'word-segmentation', 'lstm-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'part-of-speech', 'lstm-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'transformer-crf', 'transformer-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'token-classification', 'transformer-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'named-entity-recognition', 'transformer-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'word-segmentation', 'transformer-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.heads.crf_head"}, "('HEADS', 'part-of-speech', 'transformer-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/heads/crf_head.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.heads.crf_head"}, "('MODELS', 'fill-mask', 'veco')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/veco/fill_mask.py", "imports": ["transformers"], "module": "modelscope.models.nlp.veco.fill_mask"}, "('MODELS', 'token-classification', 'veco')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/veco/token_classification.py", "imports": ["transformers", "torch"], "module": "modelscope.models.nlp.veco.token_classification"}, "('MODELS', 'backbone', 'veco')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/veco/backbone.py", "imports": ["transformers"], "module": "modelscope.models.nlp.veco.backbone"}, "('MODELS', 'nli', 'veco')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/veco/text_classification.py", "imports": ["transformers"], "module": "modelscope.models.nlp.veco.text_classification"}, "('MODELS', 'sentiment-classification', 'veco')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/veco/text_classification.py", "imports": ["transformers"], "module": "modelscope.models.nlp.veco.text_classification"}, "('MODELS', 'sentence-similarity', 'veco')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/veco/text_classification.py", "imports": ["transformers"], "module": "modelscope.models.nlp.veco.text_classification"}, "('MODELS', 'text-classification', 'veco')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/veco/text_classification.py", "imports": ["transformers"], "module": "modelscope.models.nlp.veco.text_classification"}, "('MODELS', 'text-generation', 'glm130b')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/glm_130b/text_generation.py", "imports": ["time", "stat", "re", "functools", "SwissArmyTransformer", "typing", "torch", "sys", "os", "copy", "random"], "module": "modelscope.models.nlp.glm_130b.text_generation"}, "('MODELS', 'backbone', 'transformers')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/hf_transformers/backbone.py", "imports": ["transformers"], "module": "modelscope.models.nlp.hf_transformers.backbone"}, "('MODELS', 'fill-mask', 'deberta_v2')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/deberta_v2/fill_mask.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.deberta_v2.fill_mask"}, "('MODELS', 'backbone', 'deberta_v2')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/deberta_v2/backbone.py", "imports": ["collections", "transformers", "typing", "torch"], "module": "modelscope.models.nlp.deberta_v2.backbone"}, "('MODELS', 'text-error-correction', 'bart')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bart/text_error_correction.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.nlp.bart.text_error_correction"}, "('MODELS', 'text-generation', 'palm-v2')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/palm_v2/text_generation.py", "imports": ["numpy", "transformers", "codecs", "dataclasses", "math", "subprocess", "json", "typing", "torch", "copy", "os"], "module": "modelscope.models.nlp.palm_v2.text_generation"}, "('MODELS', 'table-question-answering', 'space-T-en')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/space_T_en/text_to_sql.py", "imports": ["text2sql_lgesql", "os", "typing", "torch"], "module": "modelscope.models.nlp.space_T_en.text_to_sql"}, "('MODELS', 'backbone', 'xlm-roberta')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/xlm_roberta/backbone.py", "imports": ["math", "packaging", "transformers", "torch"], "module": "modelscope.models.nlp.xlm_roberta.backbone"}, "('BACKBONES', 'backbone', 'gpt2')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/gpt2/backbone.py", "imports": ["transformers"], "module": "modelscope.models.nlp.gpt2.backbone"}, "('MODELS', 'code-translation', 'codegeex')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/codegeex/codegeex_for_code_translation.py", "imports": ["copy", "typing", "torch"], "module": "modelscope.models.nlp.codegeex.codegeex_for_code_translation"}, "('MODELS', 'code-generation', 'codegeex')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/codegeex/codegeex_for_code_generation.py", "imports": ["copy", "typing", "torch"], "module": "modelscope.models.nlp.codegeex.codegeex_for_code_generation"}, "('MODELS', 'translation-evaluation', 'unite')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/unite/translation_evaluation.py", "imports": ["numpy", "transformers", "dataclasses", "math", "packaging", "warnings", "torch", "typing"], "module": "modelscope.models.nlp.unite.translation_evaluation"}, "('MODELS', 'table-question-answering', 'space-T-cn')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/space_T_cn/table_question_answering.py", "imports": ["numpy", "transformers", "typing", "torch", "os"], "module": "modelscope.models.nlp.space_T_cn.table_question_answering"}, "('MODELS', 'fill-mask', 'megatron-bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/megatron_bert/fill_mask.py", "imports": ["transformers", "torch"], "module": "modelscope.models.nlp.megatron_bert.fill_mask"}, "('MODELS', 'backbone', 'megatron-bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/megatron_bert/backbone.py", "imports": ["math", "transformers", "torch"], "module": "modelscope.models.nlp.megatron_bert.backbone"}, "('MODELS', 'backbone', 'llama2')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/llama/backbone.py", "imports": ["transformers"], "module": "modelscope.models.nlp.llama.backbone"}, "('MODELS', 'backbone', 'llama')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/llama/backbone.py", "imports": ["transformers"], "module": "modelscope.models.nlp.llama.backbone"}, "('MODELS', 'text-generation', 'llama2')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/llama/text_generation.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.llama.text_generation"}, "('MODELS', 'chat', 'llama2')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/llama/text_generation.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.llama.text_generation"}, "('MODELS', 'text-generation', 'llama')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/llama/text_generation.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.llama.text_generation"}, "('MODELS', 'text-generation', 'polylm')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/polylm/text_generation.py", "imports": ["collections", "transformers", "typing", "torch"], "module": "modelscope.models.nlp.polylm.text_generation"}, "('MODELS', 'text-generation', 'gpt-moe')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/gpt_moe/text_generation.py", "imports": ["transformers", "typing"], "module": "modelscope.models.nlp.gpt_moe.text_generation"}, "('MODELS', 'task-oriented-conversation', 'space-modeling')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/space/dialog_modeling.py", "imports": ["os", "typing"], "module": "modelscope.models.nlp.space.dialog_modeling"}, "('MODELS', 'task-oriented-conversation', 'space-dst')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/space/dialog_state_tracking.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.models.nlp.space.dialog_state_tracking"}, "('MODELS', 'task-oriented-conversation', 'space-intent')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/space/dialog_intent_prediction.py", "imports": ["os", "typing"], "module": "modelscope.models.nlp.space.dialog_intent_prediction"}, "('MODELS', 'fill-mask', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bert/fill_mask.py", "imports": [], "module": "modelscope.models.nlp.bert.fill_mask"}, "('MODELS', 'document-segmentation', 'bert-for-document-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bert/document_segmentation.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.bert.document_segmentation"}, "('MODELS', 'sentence-embedding', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bert/sentence_embedding.py", "imports": ["torch"], "module": "modelscope.models.nlp.bert.sentence_embedding"}, "('MODELS', 'siamese-uie', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bert/siamese_uie.py", "imports": ["torch", "copy"], "module": "modelscope.models.nlp.bert.siamese_uie"}, "('MODELS', 'token-classification', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bert/token_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.token_classification"}, "('MODELS', 'part-of-speech', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bert/token_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.token_classification"}, "('MODELS', 'word-segmentation', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bert/token_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.token_classification"}, "('MODELS', 'backbone', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bert/backbone.py", "imports": ["math", "packaging", "transformers", "torch"], "module": "modelscope.models.nlp.bert.backbone"}, "('MODELS', 'word-alignment', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bert/word_alignment.py", "imports": ["torch"], "module": "modelscope.models.nlp.bert.word_alignment"}, "('MODELS', 'text-ranking', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bert/text_ranking.py", "imports": [], "module": "modelscope.models.nlp.bert.text_ranking"}, "('MODELS', 'text-classification', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bert/text_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.text_classification"}, "('MODELS', 'nli', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bert/text_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.text_classification"}, "('MODELS', 'sentiment-classification', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bert/text_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.text_classification"}, "('MODELS', 'sentence-similarity', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bert/text_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.text_classification"}, "('MODELS', 'zero-shot-classification', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bert/text_classification.py", "imports": [], "module": "modelscope.models.nlp.bert.text_classification"}, "('BACKBONES', 'backbone', 'bloom')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bloom/backbone.py", "imports": ["transformers"], "module": "modelscope.models.nlp.bloom.backbone"}, "('MODELS', 'text-generation', 'bloom')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/bloom/text_generation.py", "imports": ["transformers"], "module": "modelscope.models.nlp.bloom.text_generation"}, "('MODELS', 'chat', 'chatglm6b')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/chatglm/text_generation.py", "imports": ["transformers", "re", "math", "warnings", "torch", "typing", "sys", "copy", "os"], "module": "modelscope.models.nlp.chatglm.text_generation"}, "('MODELS', 'fid-dialogue', 'fid-T5')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/fid_T5/text_generation.py", "imports": ["io", "transformers", "os", "torch"], "module": "modelscope.models.nlp.fid_T5.text_generation"}, "('MODELS', 'text-classification', 'peer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/peer/text_classification.py", "imports": ["copy", "torch"], "module": "modelscope.models.nlp.peer.text_classification"}, "('MODELS', 'nli', 'peer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/peer/text_classification.py", "imports": ["copy", "torch"], "module": "modelscope.models.nlp.peer.text_classification"}, "('MODELS', 'sentiment-classification', 'peer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/peer/text_classification.py", "imports": ["copy", "torch"], "module": "modelscope.models.nlp.peer.text_classification"}, "('MODELS', 'sentence-similarity', 'peer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/peer/text_classification.py", "imports": ["copy", "torch"], "module": "modelscope.models.nlp.peer.text_classification"}, "('MODELS', 'zero-shot-classification', 'peer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/peer/text_classification.py", "imports": ["copy", "torch"], "module": "modelscope.models.nlp.peer.text_classification"}, "('MODELS', 'competency-aware-translation', 'canmt')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/canmt/canmt_translation.py", "imports": ["numpy", "math", "typing", "torch", "os"], "module": "modelscope.models.nlp.canmt.canmt_translation"}, "('MODELS', 'backbone', 'qwen-7b')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/qwen/backbone.py", "imports": ["transformers", "math", "typing", "torch", "einops", "importlib", "flash_attn"], "module": "modelscope.models.nlp.qwen.backbone"}, "('MODELS', 'text-generation', 'qwen-7b')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/qwen/text_generation.py", "imports": ["warnings", "transformers", "typing", "torch"], "module": "modelscope.models.nlp.qwen.text_generation"}, "('MODELS', 'chat', 'qwen-7b')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/qwen/text_generation.py", "imports": ["warnings", "transformers", "typing", "torch"], "module": "modelscope.models.nlp.qwen.text_generation"}, "('MODELS', 'fid-dialogue', 'fid-plug')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/fid_plug/text_generation.py", "imports": ["io", "transformers", "os", "torch"], "module": "modelscope.models.nlp.fid_plug.text_generation"}, "('MODELS', 'text-classification', 'user-satisfaction-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/use/user_satisfaction_estimation.py", "imports": ["numpy", "transformers", "typing", "torch", "os"], "module": "modelscope.models.nlp.use.user_satisfaction_estimation"}, "('BACKBONES', 'backbone', 'gpt-neo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/gpt_neo/backbone.py", "imports": ["transformers"], "module": "modelscope.models.nlp.gpt_neo.backbone"}, "('MODELS', 'text-summarization', 'mglm')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/mglm/mglm_for_text_summarization.py", "imports": ["numpy", "megatron_util", "typing", "torch", "random", "os"], "module": "modelscope.models.nlp.mglm.mglm_for_text_summarization"}, "('MODELS', 'token-classification', 'lstm-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/lstm/token_classification.py", "imports": [], "module": "modelscope.models.nlp.lstm.token_classification"}, "('MODELS', 'named-entity-recognition', 'lstm-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/lstm/token_classification.py", "imports": [], "module": "modelscope.models.nlp.lstm.token_classification"}, "('MODELS', 'part-of-speech', 'lstm-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/lstm/token_classification.py", "imports": [], "module": "modelscope.models.nlp.lstm.token_classification"}, "('MODELS', 'word-segmentation', 'lstm-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/lstm/token_classification.py", "imports": [], "module": "modelscope.models.nlp.lstm.token_classification"}, "('MODELS', 'word-segmentation', 'lstm-crf-for-word-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/lstm/token_classification.py", "imports": [], "module": "modelscope.models.nlp.lstm.token_classification"}, "('MODELS', 'backbone', 'lstm')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/lstm/backbone.py", "imports": ["torch"], "module": "modelscope.models.nlp.lstm.backbone"}, "('MODELS', 'document-grounded-dialog-retrieval', 'doc2bot')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/dgds/document_grounded_dialog_retrieval.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.nlp.dgds.document_grounded_dialog_retrieval"}, "('MODELS', 'document-grounded-dialog-rerank', 'doc2bot')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/dgds/document_grounded_dialog_rerank.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.nlp.dgds.document_grounded_dialog_rerank"}, "('MODELS', 'document-grounded-dialog-generate', 'doc2bot')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/dgds/document_grounded_dialog_generate.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.nlp.dgds.document_grounded_dialog_generate"}, "('MODELS', 'fill-mask', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/structbert/fill_mask.py", "imports": ["transformers", "torch"], "module": "modelscope.models.nlp.structbert.fill_mask"}, "('MODELS', 'faq-question-answering', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/structbert/faq_question_answering.py", "imports": ["collections", "math", "typing", "torch", "os"], "module": "modelscope.models.nlp.structbert.faq_question_answering"}, "('MODELS', 'token-classification', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/structbert/token_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.token_classification"}, "('MODELS', 'word-segmentation', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/structbert/token_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.token_classification"}, "('MODELS', 'part-of-speech', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/structbert/token_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.token_classification"}, "('MODELS', 'backbone', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/structbert/backbone.py", "imports": ["transformers", "dataclasses", "math", "packaging", "typing", "torch"], "module": "modelscope.models.nlp.structbert.backbone"}, "('MODELS', 'text-classification', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/structbert/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.text_classification"}, "('MODELS', 'nli', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/structbert/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.text_classification"}, "('MODELS', 'sentiment-classification', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/structbert/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.text_classification"}, "('MODELS', 'sentence-similarity', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/structbert/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.text_classification"}, "('MODELS', 'zero-shot-classification', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/structbert/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.structbert.text_classification"}, "('MODELS', 'backbone', 'plug-mental')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/plug_mental/backbone.py", "imports": ["transformers", "dataclasses", "math", "packaging", "typing", "torch"], "module": "modelscope.models.nlp.plug_mental.backbone"}, "('MODELS', 'text-classification', 'plug-mental')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/plug_mental/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.plug_mental.text_classification"}, "('MODELS', 'nli', 'plug-mental')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/plug_mental/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.plug_mental.text_classification"}, "('MODELS', 'sentiment-classification', 'plug-mental')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/plug_mental/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.plug_mental.text_classification"}, "('MODELS', 'sentence-similarity', 'plug-mental')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/plug_mental/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.plug_mental.text_classification"}, "('MODELS', 'zero-shot-classification', 'plug-mental')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/plug_mental/text_classification.py", "imports": ["torch"], "module": "modelscope.models.nlp.plug_mental.text_classification"}, "('MODELS', 'text-generation', 'gpt3')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/gpt3/text_generation.py", "imports": ["collections", "transformers", "typing", "torch"], "module": "modelscope.models.nlp.gpt3.text_generation"}, "('MODELS', 'chat', 'chatglm2-6b')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/chatglm2/text_generation.py", "imports": ["transformers", "math", "typing", "warnings", "torch", "sys", "copy"], "module": "modelscope.models.nlp.chatglm2.text_generation"}, "('MODELS', 'translation', 'csanmt-translation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/csanmt/translation.py", "imports": ["tensorflow", "math", "collections", "typing"], "module": "modelscope.models.nlp.csanmt.translation"}, "('MODELS', 'fill-mask', 'fill-mask')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/fill_mask.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.models.nlp.task_models.fill_mask"}, "('MODELS', 'information-extraction', 'information-extraction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/information_extraction.py", "imports": ["numpy", "typing"], "module": "modelscope.models.nlp.task_models.information_extraction"}, "('MODELS', 'relation-extraction', 'information-extraction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/information_extraction.py", "imports": ["numpy", "typing"], "module": "modelscope.models.nlp.task_models.information_extraction"}, "('MODELS', 'token-classification', 'token-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'part-of-speech', 'token-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'named-entity-recognition', 'token-classification-for-ner')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'transformer-crf', 'transformer-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'token-classification', 'transformer-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'token-classification', 'transformer-crf-for-word-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'named-entity-recognition', 'transformer-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'part-of-speech', 'transformer-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'word-segmentation', 'transformer-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'word-segmentation', 'transformer-crf-for-word-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/token_classification.py", "imports": ["typing", "torch"], "module": "modelscope.models.nlp.task_models.token_classification"}, "('MODELS', 'machine-reading-comprehension', 'machine-reading-comprehension')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/machine_reading_comprehension.py", "imports": ["transformers", "dataclasses", "typing", "torch", "os"], "module": "modelscope.models.nlp.task_models.machine_reading_comprehension"}, "('MODELS', 'text-generation', 'text-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/text_generation.py", "imports": ["numpy", "transformers", "typing", "torch"], "module": "modelscope.models.nlp.task_models.text_generation"}, "('MODELS', 'text-ranking', 'text-ranking')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/text_ranking.py", "imports": ["numpy", "typing"], "module": "modelscope.models.nlp.task_models.text_ranking"}, "('MODELS', 'text-classification', 'text-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/text_classification.py", "imports": ["numpy", "typing"], "module": "modelscope.models.nlp.task_models.text_classification"}, "('MODELS', 'feature-extraction', 'feature-extraction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/task_models/feature_extraction.py", "imports": ["numpy", "typing"], "module": "modelscope.models.nlp.task_models.feature_extraction"}, "('MODELS', 'text2text-generation', 'T5')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/T5/text2text_generation.py", "imports": ["transformers", "warnings", "typing", "torch", "copy"], "module": "modelscope.models.nlp.T5.text2text_generation"}, "('MODELS', 'backbone', 'T5')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/nlp/T5/backbone.py", "imports": ["transformers", "math", "typing", "warnings", "torch", "copy", "os"], "module": "modelscope.models.nlp.T5.backbone"}, "('MODELS', 'protein-structure', 'unifold')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/science/unifold/model.py", "imports": ["os", "argparse", "typing", "torch"], "module": "modelscope.models.science.unifold.model"}, "('MODELS', 'punctuation', 'generic-punc')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/punc/generic_punctuation.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.punc.generic_punctuation"}, "('MODELS', 'inverse-text-processing', 'generic-itn')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/itn/generic_inverse_text_processing.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.itn.generic_inverse_text_processing"}, "('MODELS', 'acoustic-noise-suppression', 'speech_dfsmn_ans')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/ans/denoise_net.py", "imports": ["torch"], "module": "modelscope.models.audio.ans.denoise_net"}, "('MODELS', 'acoustic-noise-suppression', 'speech_frcrn_ans_cirm_16k')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/ans/frcrn.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.audio.ans.frcrn"}, "('MODELS', 'speech-separation', 'speech_mossformer2_separation_temporal_8k')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/separation/m2/mossformer.py", "imports": ["os", "torch"], "module": "modelscope.models.audio.separation.m2.mossformer"}, "('MODELS', 'speech-separation', 'speech_mossformer_separation_temporal_8k')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/separation/mossformer.py", "imports": ["copy", "os", "typing", "torch"], "module": "modelscope.models.audio.separation.mossformer"}, "('MODELS', 'speaker-verification', 'rdino_ecapa-tdnn-sv')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/rdino.py", "imports": ["math", "typing", "torch", "os", "torchaudio"], "module": "modelscope.models.audio.sv.rdino"}, "('HEADS', 'speaker-diarization-semantic-speaker-turn-detection', 'token-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/speaker_diarization_semantic_speaker_turn_detection.py", "imports": ["torch"], "module": "modelscope.models.audio.sv.speaker_diarization_semantic_speaker_turn_detection"}, "('MODELS', 'speaker-diarization-semantic-speaker-turn-detection', 'token-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/speaker_diarization_semantic_speaker_turn_detection.py", "imports": ["torch"], "module": "modelscope.models.audio.sv.speaker_diarization_semantic_speaker_turn_detection"}, "('MODELS', 'speaker-diarization-semantic-speaker-turn-detection', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/speaker_diarization_semantic_speaker_turn_detection.py", "imports": ["torch"], "module": "modelscope.models.audio.sv.speaker_diarization_semantic_speaker_turn_detection"}, "('MODELS', 'speaker-verification', 'cam++-sv')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/DTDNN.py", "imports": ["numpy", "collections", "typing", "torch", "os", "torchaudio"], "module": "modelscope.models.audio.sv.DTDNN"}, "('MODELS', 'speaker-verification', 'eres2net-sv')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/ERes2Net.py", "imports": ["math", "typing", "torch", "os", "torchaudio"], "module": "modelscope.models.audio.sv.ERes2Net"}, "('MODELS', 'speaker-verification', 'eres2net-aug-sv')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/ERes2Net_aug.py", "imports": ["math", "typing", "torch", "os", "torchaudio"], "module": "modelscope.models.audio.sv.ERes2Net_aug"}, "('MODELS', 'speaker-diarization', 'cluster-backend')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/cluster_backend.py", "imports": ["numpy", "umap", "typing", "sklearn", "hdbscan", "scipy"], "module": "modelscope.models.audio.sv.cluster_backend"}, "('MODELS', 'speaker-verification', 'ecapa-tdnn-sv')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/ecapa_tdnn.py", "imports": ["numpy", "math", "typing", "torch", "os", "torchaudio"], "module": "modelscope.models.audio.sv.ecapa_tdnn"}, "('MODELS', 'speaker-verification', 'generic-sv')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/generic_speaker_verification.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.sv.generic_speaker_verification"}, "('MODELS', 'speaker-diarization', 'generic-sv')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/generic_speaker_verification.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.sv.generic_speaker_verification"}, "('MODELS', 'speaker-diarization', 'scl-sd')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/speaker_change_locator.py", "imports": ["numpy", "collections", "typing", "torch", "os", "torchaudio"], "module": "modelscope.models.audio.sv.speaker_change_locator"}, "('HEADS', 'speaker-diarization-dialogue-detection', 'text-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/speaker_diarization_dialogue_detection.py", "imports": ["torch"], "module": "modelscope.models.audio.sv.speaker_diarization_dialogue_detection"}, "('MODELS', 'speaker-diarization-dialogue-detection', 'text-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/speaker_diarization_dialogue_detection.py", "imports": ["torch"], "module": "modelscope.models.audio.sv.speaker_diarization_dialogue_detection"}, "('MODELS', 'speaker-diarization-dialogue-detection', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/speaker_diarization_dialogue_detection.py", "imports": ["torch"], "module": "modelscope.models.audio.sv.speaker_diarization_dialogue_detection"}, "('MODELS', 'speech-language-recognition', 'eres2net-lre')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/lanuage_recognition_eres2net.py", "imports": ["numpy", "typing", "torch", "os", "torchaudio"], "module": "modelscope.models.audio.sv.lanuage_recognition_eres2net"}, "('MODELS', 'speech-language-recognition', 'cam++-lre')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/sv/lanuage_recognition_model.py", "imports": ["numpy", "typing", "torch", "os", "torchaudio"], "module": "modelscope.models.audio.sv.lanuage_recognition_model"}, "('MODELS', 'text-to-speech', 'sambert-hifigan')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/tts/sambert_hifi.py", "imports": ["wave", "numpy", "__future__", "yaml", "json", "shutil", "matplotlib", "datetime", "zipfile", "os"], "module": "modelscope.models.audio.tts.sambert_hifi"}, "('MODELS', 'keyword-spotting', 'speech_dfsmn_kws_char_farfield')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/kws/farfield/model.py", "imports": ["tempfile", "os", "typing"], "module": "modelscope.models.audio.kws.farfield.model"}, "('MODELS', 'keyword-spotting', 'speech_dfsmn_kws_char_farfield_iot')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/kws/farfield/model.py", "imports": ["tempfile", "os", "typing"], "module": "modelscope.models.audio.kws.farfield.model"}, "('MODELS', 'keyword-spotting', 'speech_kws_fsmn_char_ctc_nearfield')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/kws/nearfield/model.py", "imports": ["typing", "torch", "tempfile", "sys", "os"], "module": "modelscope.models.audio.kws.nearfield.model"}, "('MODELS', 'keyword-spotting', 'kws-kwsbp')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/kws/generic_key_word_spotting.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.kws.generic_key_word_spotting"}, "('MODELS', 'auto-speech-recognition', 'generic-asr')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/asr/generic_automatic_speech_recognition.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.asr.generic_automatic_speech_recognition"}, "('MODELS', 'voice-activity-detection', 'generic-asr')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/asr/generic_automatic_speech_recognition.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.asr.generic_automatic_speech_recognition"}, "('MODELS', 'speech-separation', 'generic-asr')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/asr/generic_automatic_speech_recognition.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.asr.generic_automatic_speech_recognition"}, "('MODELS', 'language-score-prediction', 'generic-asr')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/asr/generic_automatic_speech_recognition.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.asr.generic_automatic_speech_recognition"}, "('MODELS', 'speech-timestamp', 'generic-asr')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/asr/generic_automatic_speech_recognition.py", "imports": ["os", "typing"], "module": "modelscope.models.audio.asr.generic_automatic_speech_recognition"}, "('MODELS', 'auto-speech-recognition', 'wenet-asr')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/audio/asr/wenet_automatic_speech_recognition.py", "imports": ["os", "json", "wenetruntime", "typing"], "module": "modelscope.models.audio.asr.wenet_automatic_speech_recognition"}, "('MODELS', 'image-colorization', 'ddcolor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_colorization/ddcolor/ddcolor_for_image_colorization.py", "imports": ["numpy", "typing", "torch", "copy", "os"], "module": "modelscope.models.cv.image_colorization.ddcolor.ddcolor_for_image_colorization"}, "('MODELS', 'face-detection', 'retinaface')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/retinaface/detection.py", "imports": ["numpy", "cv2", "torch"], "module": "modelscope.models.cv.face_detection.retinaface.detection"}, "('MODELS', 'face-detection', 'ulfd')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/ulfd_slim/detection.py", "imports": ["numpy", "cv2", "os", "torch"], "module": "modelscope.models.cv.face_detection.ulfd_slim.detection"}, "('PREPROCESSORS', 'cv', 'object-detection-scrfd')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/preprocessor.py", "imports": ["PIL", "numpy", "typing"], "module": "modelscope.models.cv.face_detection.scrfd.preprocessor"}, "('DETECTORS', 'default', 'TinyMog')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/detectors/tinymog.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.detectors.tinymog"}, "('DETECTORS', 'default', 'SCRFD')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/detectors/scrfd.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.detectors.scrfd"}, "('DETECTORS', 'default', 'CustomSingleStageDetector')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/detectors/single_stage.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.detectors.single_stage"}, "('HEADS', 'default', 'SCRFDHead')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/dense_heads/scrfd_head.py", "imports": ["mmcv", "numpy", "mmdet", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.dense_heads.scrfd_head"}, "('BACKBONES', 'default', 'MasterNet')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/backbones/master_net.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.backbones.master_net"}, "('BACKBONES', 'default', 'ResNetV1e')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/backbones/resnet.py", "imports": ["mmcv", "mmdet", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.backbones.resnet"}, "('BACKBONES', 'default', 'MobileNetV1')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/models/backbones/mobilenet.py", "imports": ["mmcv", "mmdet", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.backbones.mobilenet"}, "('PIPELINES', 'default', 'ResizeV2')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/transforms.py", "imports": ["mmcv", "mmdet", "numpy"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.transforms"}, "('PIPELINES', 'default', 'RandomFlipV2')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/transforms.py", "imports": ["mmcv", "mmdet", "numpy"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.transforms"}, "('PIPELINES', 'default', 'RandomSquareCrop')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/transforms.py", "imports": ["mmcv", "mmdet", "numpy"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.transforms"}, "('PIPELINES', 'default', 'DefaultFormatBundleV2')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/formating.py", "imports": ["mmcv", "numpy", "mmdet", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.formating"}, "('PIPELINES', 'default', 'LoadAnnotationsV2')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/loading.py", "imports": ["numpy", "mmdet", "os", "pycocotools"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.loading"}, "('PIPELINES', 'default', 'RotateV2')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/pipelines/auto_augment.py", "imports": ["numpy", "copy", "mmcv", "mmdet", "cv2"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.auto_augment"}, "('DATASETS', 'default', 'RetinaFaceDataset')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/mmdet_patch/datasets/retinaface.py", "imports": ["numpy", "mmdet"], "module": "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.retinaface"}, "('MODELS', 'face-detection', 'tinymog')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/tinymog_detect.py", "imports": ["copy", "os", "typing", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.tinymog_detect"}, "('MODELS', 'face-detection', 'scrfd')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/scrfd_detect.py", "imports": ["numpy", "typing", "torch", "copy", "os"], "module": "modelscope.models.cv.face_detection.scrfd.scrfd_detect"}, "('MODELS', 'card-detection', 'scrfd')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/scrfd_detect.py", "imports": ["numpy", "typing", "torch", "copy", "os"], "module": "modelscope.models.cv.face_detection.scrfd.scrfd_detect"}, "('MODELS', 'face-detection', 'damofd')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/scrfd/damofd_detect.py", "imports": ["copy", "os", "typing", "torch"], "module": "modelscope.models.cv.face_detection.scrfd.damofd_detect"}, "('MODELS', 'face-detection', 'mtcnn')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/mtcnn/models/detector.py", "imports": ["PIL", "numpy", "os", "torch"], "module": "modelscope.models.cv.face_detection.mtcnn.models.detector"}, "('MODELS', 'face-detection', 'mogface')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_detection/mogface/models/detectors.py", "imports": ["numpy", "cv2", "os", "torch"], "module": "modelscope.models.cv.face_detection.mogface.models.detectors"}, "('MODELS', 'product-segmentation', 'product-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/product_segmentation/seg_infer.py", "imports": ["PIL", "numpy", "cv2", "torch"], "module": "modelscope.models.cv.product_segmentation.seg_infer"}, "('MODELS', 'video-summarization', 'pgl-video-summarization')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_summarization/summarizer.py", "imports": ["numpy", "os", "typing", "torch"], "module": "modelscope.models.cv.video_summarization.summarizer"}, "('PREPROCESSORS', 'cv', 'image-sky-change-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_skychange/preprocessor.py", "imports": ["numpy", "torchvision", "json", "numbers", "typing", "cv2", "pdb", "torch"], "module": "modelscope.models.cv.image_skychange.preprocessor"}, "('MODELS', 'image-skychange', 'image-skychange')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_skychange/skychange_model.py", "imports": ["collections", "time", "math", "json", "typing", "cv2", "pdb", "torch", "os"], "module": "modelscope.models.cv.image_skychange.skychange_model"}, "('MODELS', 'image-object-detection', 'tinynas-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/tinynas_detection/tinynas_detector.py", "imports": [], "module": "modelscope.models.cv.tinynas_detection.tinynas_detector"}, "('MODELS', 'domain-specific-object-detection', 'tinynas-damoyolo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/tinynas_detection/tinynas_damoyolo.py", "imports": [], "module": "modelscope.models.cv.tinynas_detection.tinynas_damoyolo"}, "('MODELS', 'image-object-detection', 'tinynas-damoyolo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/tinynas_detection/tinynas_damoyolo.py", "imports": [], "module": "modelscope.models.cv.tinynas_detection.tinynas_damoyolo"}, "('MODELS', 'face-emotion', 'face-emotion')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_emotion/emotion_model.py", "imports": ["os", "sys", "torch"], "module": "modelscope.models.cv.face_emotion.emotion_model"}, "('MODELS', 'image-quality-assessment-mos', 'image-quality-assessment-mos')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_quality_assessment_mos/image_quality_assessment_mos.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.image_quality_assessment_mos.image_quality_assessment_mos"}, "('MODELS', 'image-quality-assessment-degradation', 'image-quality-assessment-degradation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_quality_assessment_degradation/image_quality_assessment_degradation.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.image_quality_assessment_degradation.image_quality_assessment_degradation"}, "('MODELS', 'pedestrian-attribute-recognition', 'pedestrian-attribute-recognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/pedestrian_attribute_recognition/model.py", "imports": ["numpy", "os", "torchvision", "torch"], "module": "modelscope.models.cv.pedestrian_attribute_recognition.model"}, "('MODELS', 'image-depth-estimation', 'newcrfs-depth-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_depth_estimation/newcrfs_model.py", "imports": ["numpy", "os", "torch"], "module": "modelscope.models.cv.image_depth_estimation.newcrfs_model"}, "('MODELS', 'panorama-depth-estimation', 'unifuse-depth-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/panorama_depth_estimation/unifuse_model.py", "imports": ["numpy", "os", "torchvision", "torch"], "module": "modelscope.models.cv.panorama_depth_estimation.unifuse_model"}, "('MODELS', 'bad-image-detecting', 'bad-image-detecting')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/bad_image_detecting/bad_image_detecting.py", "imports": ["numpy", "torchvision", "typing", "torch", "os"], "module": "modelscope.models.cv.bad_image_detecting.bad_image_detecting"}, "('MODELS', 'video-super-resolution', 'real-basicvsr')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_super_resolution/real_basicvsr_for_video_super_resolution.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.video_super_resolution.real_basicvsr_for_video_super_resolution"}, "('MODELS', 'video-super-resolution', 'msrresnet-lite')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_super_resolution/msrresnet_lite_model.py", "imports": ["functools", "os", "typing", "torch"], "module": "modelscope.models.cv.video_super_resolution.msrresnet_lite_model"}, "('PREPROCESSORS', 'cv', 'ocr-recognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/ocr_recognition/preprocessor.py", "imports": ["numpy", "cv2", "PIL", "torch", "os"], "module": "modelscope.models.cv.ocr_recognition.preprocessor"}, "('MODELS', 'ocr-recognition', 'OCRRecognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/ocr_recognition/model.py", "imports": ["os", "torch"], "module": "modelscope.models.cv.ocr_recognition.model"}, "('MODELS', 'image-debanding', 'rrdb')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_debanding/rrdb/rrdb_image_debanding.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.image_debanding.rrdb.rrdb_image_debanding"}, "('MODELS', 'crowd-counting', 'HRNetCrowdCounting')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/crowd_counting/cc_model.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.crowd_counting.cc_model"}, "('MODELS', 'video-frame-interpolation', 'video-frame-interpolation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_frame_interpolation/VFINet_for_video_frame_interpolation.py", "imports": ["copy", "os", "typing", "torch"], "module": "modelscope.models.cv.video_frame_interpolation.VFINet_for_video_frame_interpolation"}, "('MODELS', 'face-attribute-recognition', 'fairface')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_attribute_recognition/fair_face/face_attribute_recognition.py", "imports": ["numpy", "torchvision", "cv2", "PIL", "torch", "os"], "module": "modelscope.models.cv.face_attribute_recognition.fair_face.face_attribute_recognition"}, "('MODELS', 'pointcloud-sceneflow-estimation', 'rcp-sceneflow-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/pointcloud_sceneflow_estimation/rcp_model.py", "imports": ["numpy", "os", "torch"], "module": "modelscope.models.cv.pointcloud_sceneflow_estimation.rcp_model"}, "('MODELS', 'image-segmentation', 'swinL-panoptic-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_panoptic_segmentation/panseg_model.py", "imports": ["os", "torch"], "module": "modelscope.models.cv.image_panoptic_segmentation.panseg_model"}, "('MODELS', 'image-try-on', 'image-try-on')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_try_on/try_on_infer.py", "imports": ["numpy", "torchvision", "yaml", "argparse", "torch", "cv2", "PIL", "os"], "module": "modelscope.models.cv.image_try_on.try_on_infer"}, "('MODELS', 'image-paintbyexample', 'Stablediffusion-Paintbyexample')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_paintbyexample/model.py", "imports": ["typing", "torch", "omegaconf", "paint_ldm", "os"], "module": "modelscope.models.cv.image_paintbyexample.model"}, "('MODELS', 'face-recognition', 'rts-backbone')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_recognition/torchkit/rts_backbone.py", "imports": ["math", "collections", "os", "torch"], "module": "modelscope.models.cv.face_recognition.torchkit.rts_backbone"}, "('MODELS', 'image-segmentation', 'maskdino_swin')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_instance_segmentation/maskdino_model.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.image_instance_segmentation.maskdino_model"}, "('MODELS', 'image-segmentation', 'fastinst')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_instance_segmentation/fastinst_model.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.image_instance_segmentation.fastinst_model"}, "('MODELS', 'image-segmentation', 'cascade_mask_rcnn_swin')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_instance_segmentation/model.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.image_instance_segmentation.model"}, "('MODELS', 'product-retrieval-embedding', 'product-retrieval-embedding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/product_retrieval_embedding/item_model.py", "imports": ["numpy", "os", "typing", "torch"], "module": "modelscope.models.cv.product_retrieval_embedding.item_model"}, "('MODELS', 'image-super-resolution', 'ecbsr')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/super_resolution/ecbsr_model.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.super_resolution.ecbsr_model"}, "('MODELS', 'surface-recon-common', 'surface-recon-common')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/surface_recon_common/surface_recon_common.py", "imports": ["numpy", "os", "trimesh", "torch"], "module": "modelscope.models.cv.surface_recon_common.surface_recon_common"}, "('MODELS', 'image-matching', 'quadtree-attention-image-matching')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_matching/quadtree_attention_model.py", "imports": ["numpy", "cv2", "torch", "pathlib", "os"], "module": "modelscope.models.cv.image_matching.quadtree_attention_model"}, "('MODELS', 'video-text-retrieval', 'vop-retrieval-model-se')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/vop_retrieval/model_se.py", "imports": ["os", "torch"], "module": "modelscope.models.cv.vop_retrieval.model_se"}, "('MODELS', 'video-text-retrieval', 'vop-retrieval-model')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/vop_retrieval/model.py", "imports": ["os", "torch"], "module": "modelscope.models.cv.vop_retrieval.model"}, "('MODELS', 'nerf-recon-4k', 'nerf-recon-4k')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/nerf_recon_4k/nerf_recon_4k.py", "imports": ["numpy", "time", "tqdm", "argparse", "torch", "random", "mmcv", "os", "imageio"], "module": "modelscope.models.cv.nerf_recon_4k.nerf_recon_4k"}, "('PREPROCESSORS', 'cv', 'nerf-recon-acc-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/nerf_recon_acc/nerf_preprocess.py", "imports": ["numpy", "subprocess", "typing", "cv2", "os", "tensorflow", "glob"], "module": "modelscope.models.cv.nerf_recon_acc.nerf_preprocess"}, "('MODELS', 'language-guided-video-summarization', 'clip-it-language-guided-video-summarization')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/language_guided_video_summarization/summarizer.py", "imports": ["numpy", "videofeatures_clipit", "argparse", "typing", "torch", "bmt_clipit", "os"], "module": "modelscope.models.cv.language_guided_video_summarization.summarizer"}, "('MODELS', 'nerf-recon-vq-compression', 'nerf-recon-vq-compression')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/nerf_recon_vq_compression/nerf_recon_vq_compression.py", "imports": ["numpy", "time", "functools", "tqdm", "cv2", "torch", "os", "glob"], "module": "modelscope.models.cv.nerf_recon_vq_compression.nerf_recon_vq_compression"}, "('MODELS', 'image-classification', 'image-probing-model')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_probing_model/model.py", "imports": ["json", "os", "typing", "torch"], "module": "modelscope.models.cv.image_probing_model.model"}, "('MODELS', 'lineless-table-recognition', 'LoreModel')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/table_recognition/model_lore.py", "imports": ["numpy", "math", "typing", "torch", "copy", "os"], "module": "modelscope.models.cv.table_recognition.model_lore"}, "('MODELS', 'video-panoptic-segmentation', 'swinb-video-panoptic-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_panoptic_segmentation/video_k_net.py", "imports": ["mmcv", "numpy", "mmdet", "torch"], "module": "modelscope.models.cv.video_panoptic_segmentation.video_k_net"}, "('TRACKERS', 'default', 'QuasiDenseEmbedTracker')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_panoptic_segmentation/track/quasi_dense_embed_tracker.py", "imports": ["mmcv", "mmdet", "torch"], "module": "modelscope.models.cv.video_panoptic_segmentation.track.quasi_dense_embed_tracker"}, "('NECKS', 'default', 'SemanticFPNWrapper')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_panoptic_segmentation/head/semantic_fpn_wrapper.py", "imports": ["mmcv", "mmdet", "torch"], "module": "modelscope.models.cv.video_panoptic_segmentation.head.semantic_fpn_wrapper"}, "('TRANSFORMER_LAYER', 'default', 'KernelUpdator')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_instance_segmentation/head/kernel_updator.py", "imports": ["mmcv", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.head.kernel_updator"}, "('HEADS', 'default', 'VideoKernelUpdateHead')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_panoptic_segmentation/head/kernel_update_head.py", "imports": ["mmcv", "numpy", "mmdet", "torch"], "module": "modelscope.models.cv.video_panoptic_segmentation.head.kernel_update_head"}, "('HEADS', 'default', 'VideoKernelIterHead')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_panoptic_segmentation/head/kernel_iter_head.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.video_panoptic_segmentation.head.kernel_iter_head"}, "('MODELS', 'face-human-hand-detection', 'face-human-hand-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_human_hand_detection/det_infer.py", "imports": ["numpy", "cv2", "torch"], "module": "modelscope.models.cv.face_human_hand_detection.det_infer"}, "('MODELS', 'image-portrait-enhancement', 'gpen')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_portrait_enhancement/image_portrait_enhancement.py", "imports": ["math", "os", "typing", "torch"], "module": "modelscope.models.cv.image_portrait_enhancement.image_portrait_enhancement"}, "('MODELS', 'image-inpainting', 'FFTInpainting')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_inpainting/model.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.image_inpainting.model"}, "('MODELS', 'open-vocabulary-detection', 'open-vocabulary-detection-vild')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/open_vocabulary_detection_vild/vild.py", "imports": ["numpy", "typing", "torch", "scipy", "clip", "os", "tensorflow"], "module": "modelscope.models.cv.open_vocabulary_detection_vild.vild"}, "('MODELS', 'image-quality-assessment-mos', 'image-quality-assessment-man')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_quality_assessment_man/image_quality_assessment_man.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.image_quality_assessment_man.image_quality_assessment_man"}, "('MODELS', 'face-reconstruction', 'face_reconstruction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/face_reconstruction/models/facerecon_model.py", "imports": ["numpy", "collections", "cv2", "torch", "os"], "module": "modelscope.models.cv.face_reconstruction.models.facerecon_model"}, "('MODELS', 'image-reid-person', 'passvitb')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_reid_person/pass_model.py", "imports": ["enum", "os", "torch"], "module": "modelscope.models.cv.image_reid_person.pass_model"}, "('MODELS', 'image-color-enhancement', 'csrnet')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_color_enhance/image_color_enhance.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.image_color_enhance.image_color_enhance"}, "('MODELS', 'image-color-enhancement', 'deeplpfnet')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_color_enhance/deeplpf/deeplpf_image_color_enhance.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.image_color_enhance.deeplpf.deeplpf_image_color_enhance"}, "('MODELS', 'image-color-enhancement', 'adaint')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_color_enhance/adaint/adaint.py", "imports": ["torchvision", "numbers", "typing", "torch", "os"], "module": "modelscope.models.cv.image_color_enhance.adaint.adaint"}, "('MODELS', 'semantic-segmentation', 'detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/salient_detection/salient_model.py", "imports": ["torchvision", "cv2", "PIL", "torch", "os"], "module": "modelscope.models.cv.salient_detection.salient_model"}, "('MODELS', 'movie-scene-segmentation', 'resnet50-bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/movie_scene_segmentation/model.py", "imports": ["numpy", "torchvision", "math", "tqdm", "typing", "torch", "PIL", "einops", "shotdetect_scenedetect_lgss", "os"], "module": "modelscope.models.cv.movie_scene_segmentation.model"}, "('MODELS', 'video-human-matting', 'video-human-matting')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_human_matting/model.py", "imports": ["numpy", "torchvision", "typing", "torch", "os"], "module": "modelscope.models.cv.video_human_matting.model"}, "('HEADS', 'default', 'RPNNHead')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection/mmdet_ms/dense_heads/rpn_head.py", "imports": ["mmcv", "mmdet", "copy", "torch"], "module": "modelscope.models.cv.object_detection.mmdet_ms.dense_heads.rpn_head"}, "('HEADS', 'default', 'AnchorNHead')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection/mmdet_ms/dense_heads/anchor_head.py", "imports": ["mmdet"], "module": "modelscope.models.cv.object_detection.mmdet_ms.dense_heads.anchor_head"}, "('HEADS', 'default', 'ConvFCBBoxNHead')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection/mmdet_ms/roi_heads/bbox_heads/convfc_bbox_head.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.object_detection.mmdet_ms.roi_heads.bbox_heads.convfc_bbox_head"}, "('HEADS', 'default', 'Shared2FCBBoxNHead')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection/mmdet_ms/roi_heads/bbox_heads/convfc_bbox_head.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.object_detection.mmdet_ms.roi_heads.bbox_heads.convfc_bbox_head"}, "('HEADS', 'default', 'Shared4Conv1FCBBoxNHead')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection/mmdet_ms/roi_heads/bbox_heads/convfc_bbox_head.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.object_detection.mmdet_ms.roi_heads.bbox_heads.convfc_bbox_head"}, "('HEADS', 'default', 'FCNMaskNHead')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection/mmdet_ms/roi_heads/mask_heads/fcn_mask_head.py", "imports": ["numpy", "mmdet", "warnings", "torch", "mmcv"], "module": "modelscope.models.cv.object_detection.mmdet_ms.roi_heads.mask_heads.fcn_mask_head"}, "('BACKBONES', 'default', 'ViT')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection/mmdet_ms/backbones/vit.py", "imports": ["functools", "math", "mmdet", "timm", "torch"], "module": "modelscope.models.cv.object_detection.mmdet_ms.backbones.vit"}, "('NECKS', 'default', 'FPNF')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection/mmdet_ms/necks/fpn.py", "imports": ["mmcv", "mmdet", "torch"], "module": "modelscope.models.cv.object_detection.mmdet_ms.necks.fpn"}, "('MODELS', 'human-detection', 'detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection/mmdet_model.py", "imports": ["numpy", "os", "torch"], "module": "modelscope.models.cv.object_detection.mmdet_model"}, "('MODELS', 'image-object-detection', 'detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection/mmdet_model.py", "imports": ["numpy", "os", "torch"], "module": "modelscope.models.cv.object_detection.mmdet_model"}, "('MODELS', 'video-object-segmentation', 'video-object-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_object_segmentation/model.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.video_object_segmentation.model"}, "('MODELS', 'image-classification', 'bnext')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_binary_quant_classification/binary_quant_model.py", "imports": ["collections", "os", "torch"], "module": "modelscope.models.cv.image_binary_quant_classification.binary_quant_model"}, "('PREPROCESSORS', 'cv', 'ocr-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/ocr_detection/preprocessor.py", "imports": ["numpy", "math", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.models.cv.ocr_detection.preprocessor"}, "('MODELS', 'ocr-detection', 'OCRDetection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/ocr_detection/model.py", "imports": ["numpy", "os", "typing", "torch"], "module": "modelscope.models.cv.ocr_detection.model"}, "('MODELS', 'human-image-generation', 'human-image-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/human_image_generation/human_image_generation_infer.py", "imports": ["numpy", "ast", "torchvision", "math", "pickle", "torch", "cv2", "PIL", "random"], "module": "modelscope.models.cv.human_image_generation.human_image_generation_infer"}, "('MODELS', 'image-object-detection', 'vidt')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/vidt/model.py", "imports": ["os", "torch"], "module": "modelscope.models.cv.vidt.model"}, "('HEADS', 'default', 'MaskFormerSemanticHead')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_semantic_segmentation/pan_merge/maskformer_semantic_head.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.image_semantic_segmentation.pan_merge.maskformer_semantic_head"}, "('MODELS', 'image-segmentation', 'swinL-semantic-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_semantic_segmentation/semantic_seg_model.py", "imports": ["numpy", "os", "torch"], "module": "modelscope.models.cv.image_semantic_segmentation.semantic_seg_model"}, "('MODELS', 'image-segmentation', 'vitadapter-semantic-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_semantic_segmentation/semantic_seg_model.py", "imports": ["numpy", "os", "torch"], "module": "modelscope.models.cv.image_semantic_segmentation.semantic_seg_model"}, "('DETECTORS', 'default', 'EncoderDecoderMask2Former')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/models/segmentors/encoder_decoder_mask2former.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.segmentors.encoder_decoder_mask2former"}, "('HEADS', 'default', 'Mask2FormerHeadFromMMSeg')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/models/decode_heads/mask2former_head_from_mmseg.py", "imports": ["mmcv", "mmdet", "copy", "torch"], "module": "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.decode_heads.mask2former_head_from_mmseg"}, "('BACKBONES', 'default', 'BEiTAdapter')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/models/backbone/beit_adapter.py", "imports": ["mmdet", "math", "timm", "torch", "logging"], "module": "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.backbone.beit_adapter"}, "('BACKBONES', 'default', 'BASEBEiT')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/models/backbone/base/beit.py", "imports": ["mmcv", "functools", "math", "mmdet", "timm", "torch"], "module": "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.backbone.base.beit"}, "('PIPELINES', 'default', 'ResizeToMultiple')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_semantic_segmentation/vit_adapter/utils/data_process_func.py", "imports": ["mmcv", "mmdet"], "module": "modelscope.models.cv.image_semantic_segmentation.vit_adapter.utils.data_process_func"}, "('MODELS', 'semantic-segmentation', 'ddpm')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_semantic_segmentation/ddpm_segmentation_model.py", "imports": ["ddpm_guided_diffusion", "os", "typing", "torch"], "module": "modelscope.models.cv.image_semantic_segmentation.ddpm_segmentation_model"}, "('MODELS', 'image-classification', 'content-check')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_classification/resnet50_cc.py", "imports": ["collections", "torchvision", "math", "torch", "os"], "module": "modelscope.models.cv.image_classification.resnet50_cc"}, "('BACKBONES', 'default', 'BEiTv2')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_classification/backbones/beit_v2.py", "imports": ["collections", "functools", "math", "warnings", "torch", "typing", "itertools", "einops", "mmcls", "mmcv", "os"], "module": "modelscope.models.cv.image_classification.backbones.beit_v2"}, "('BACKBONES', 'default', 'NextViT')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_classification/backbones/nextvit.py", "imports": ["collections", "functools", "math", "warnings", "torch", "typing", "itertools", "einops", "mmcls", "mmcv", "os"], "module": "modelscope.models.cv.image_classification.backbones.nextvit"}, "('MODELS', 'image-classification', 'ClassificationModel')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_classification/mmcls_model.py", "imports": ["os"], "module": "modelscope.models.cv.image_classification.mmcls_model"}, "('MODELS', 'image-denoising', 'nafnet')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_denoise/nafnet_for_image_denoise.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.image_denoise.nafnet_for_image_denoise"}, "('MODELS', 'vision-efficient-tuning', 'vision-efficient-tuning')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/vision_efficient_tuning/model.py", "imports": ["typing", "torch"], "module": "modelscope.models.cv.vision_efficient_tuning.model"}, "('MODELS', 'body-2d-keypoints', 'body-2d-keypoints')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/body_2d_keypoints/hrnet_v2.py", "imports": ["numpy", "os", "torch"], "module": "modelscope.models.cv.body_2d_keypoints.hrnet_v2"}, "('MODELS', 'controllable-image-generation', 'controllable-image-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/controllable_image_generation/controlnet.py", "imports": ["numpy", "control_ldm", "math", "typing", "cv2", "torch", "PIL", "tempfile", "sys", "einops", "os", "random"], "module": "modelscope.models.cv.controllable_image_generation.controlnet"}, "('MODELS', 'video-inpainting', 'video-inpainting')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_inpainting/inpainting_model.py", "imports": ["numpy", "math", "torchvision", "torch"], "module": "modelscope.models.cv.video_inpainting.inpainting_model"}, "('MODELS', 'image-body-reshaping', 'image-body-reshaping')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_body_reshaping/image_body_reshaping.py", "imports": ["numpy", "typing", "cv2", "torch", "os"], "module": "modelscope.models.cv.image_body_reshaping.image_body_reshaping"}, "('NECKS', 'default', 'MSDeformAttnPixelDecoder')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_instance_segmentation/neck/msdeformattn_decoder.py", "imports": ["mmcv", "mmdet", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.neck.msdeformattn_decoder"}, "('HEADS', 'default', 'KernelUpdateHeadVideo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_instance_segmentation/track/kernel_update_head.py", "imports": ["mmcv", "numpy", "mmdet", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.track.kernel_update_head"}, "('MATCH_COST', 'default', 'MaskCost')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_instance_segmentation/track/mask_hungarian_assigner.py", "imports": ["numpy", "mmdet", "scipy", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.track.mask_hungarian_assigner"}, "('BBOX_ASSIGNERS', 'default', 'MaskHungarianAssignerVideo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_instance_segmentation/track/mask_hungarian_assigner.py", "imports": ["numpy", "mmdet", "scipy", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.track.mask_hungarian_assigner"}, "('MODELS', 'video-instance-segmentation', 'swinb-video-instance-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_instance_segmentation/video_knet.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.video_knet"}, "('HEADS', 'default', 'ConvKernelHeadVideo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_instance_segmentation/head/kernel_head.py", "imports": ["mmcv", "mmdet", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.head.kernel_head"}, "('HEADS', 'default', 'KernelFrameIterHeadVideo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_instance_segmentation/head/kernel_frame_iter_head.py", "imports": ["mmcv", "mmdet", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.head.kernel_frame_iter_head"}, "('HEADS', 'default', 'KernelUpdateHead')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_instance_segmentation/head/kernel_update_head.py", "imports": ["mmcv", "numpy", "mmdet", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.head.kernel_update_head"}, "('HEADS', 'default', 'KernelIterHeadVideo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_instance_segmentation/head/kernel_iter_head.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.video_instance_segmentation.head.kernel_iter_head"}, "('MODELS', 'image-segmentation', 'vision-middleware')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/vision_middleware/model.py", "imports": ["json", "os", "typing", "torch"], "module": "modelscope.models.cv.vision_middleware.model"}, "('MODELS', 'image-fewshot-detection', 'defrcn')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_defrcn_fewshot/defrcn_for_fewshot.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.image_defrcn_fewshot.defrcn_for_fewshot"}, "('MODELS', 'video-object-detection', 'realtime-video-object-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/stream_yolo/realtime_video_detector.py", "imports": ["numpy", "time", "tqdm", "json", "argparse", "cv2", "torch", "logging", "os"], "module": "modelscope.models.cv.stream_yolo.realtime_video_detector"}, "('ROI_EXTRACTORS', 'default', 'SingleRoINExtractor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/abnormal_object_detection/mmdet_ms/roi_head/roi_extractors/single_level_roi_extractor.py", "imports": ["mmcv", "mmdet", "torch"], "module": "modelscope.models.cv.abnormal_object_detection.mmdet_ms.roi_head.roi_extractors.single_level_roi_extractor"}, "('HEADS', 'default', 'MaskScoringNRoIHead')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/abnormal_object_detection/mmdet_ms/roi_head/mask_scoring_roi_head.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.abnormal_object_detection.mmdet_ms.roi_head.mask_scoring_roi_head"}, "('MODELS', 'image-object-detection', 'MaskScoring')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/abnormal_object_detection/mmdet_model.py", "imports": ["numpy", "os", "torch"], "module": "modelscope.models.cv.abnormal_object_detection.mmdet_model"}, "('MODELS', 'image-demoireing', 'image-restoration')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_restoration/image_restoration_model.py", "imports": ["numpy", "cv2", "os", "torch"], "module": "modelscope.models.cv.image_restoration.image_restoration_model"}, "('MODELS', 'hand-static', 'hand-static')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/hand_static/hand_model.py", "imports": ["numpy", "torchvision", "cv2", "PIL", "torch", "sys", "os"], "module": "modelscope.models.cv.hand_static.hand_model"}, "('MODELS', 'indoor-layout-estimation', 'panovit-layout-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/indoor_layout_estimation/panovit.py", "imports": ["yacs", "numpy", "os", "torch"], "module": "modelscope.models.cv.indoor_layout_estimation.panovit"}, "('MODELS', 'nerf-recon-acc', 'nerf-recon-acc')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/nerf_recon_acc/nerf_recon_acc.py", "imports": ["numpy", "time", "tqdm", "cv2", "torch", "os", "glob"], "module": "modelscope.models.cv.nerf_recon_acc.nerf_recon_acc"}, "('MODELS', 'image-multi-view-depth-estimation', 'image-casmvs-depth-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_mvs_depth_estimation/casmvs_model.py", "imports": ["numpy", "cv2", "torch", "easydict", "os"], "module": "modelscope.models.cv.image_mvs_depth_estimation.casmvs_model"}, "('MODELS', 'referring-video-object-segmentation', 'swinT-referring-video-object-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/referring_video_object_segmentation/model.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.referring_video_object_segmentation.model"}, "('MODELS', 'human-reconstruction', 'human-reconstruction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/human_reconstruction/Reconstruction.py", "imports": ["numpy", "torchvision", "typing", "torch", "PIL", "cv2", "skimage", "os"], "module": "modelscope.models.cv.human_reconstruction.Reconstruction"}, "('MODELS', 'image-depth-estimation', 'bts-depth-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_depth_estimation_bts/depth_estimation_bts_model.py", "imports": ["os", "torch"], "module": "modelscope.models.cv.image_depth_estimation_bts.depth_estimation_bts_model"}, "('PREPROCESSORS', 'cv', 'image-driving-perception-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_driving_perception/preprocessor.py", "imports": ["numpy", "cv2", "typing", "torch"], "module": "modelscope.models.cv.image_driving_perception.preprocessor"}, "('MODELS', 'image-driving-perception', 'yolopv2')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_driving_perception/image_driving_percetion_model.py", "imports": ["numpy", "typing", "cv2", "torch", "os"], "module": "modelscope.models.cv.image_driving_perception.image_driving_percetion_model"}, "('MODELS', 'facial-expression-recognition', 'fer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/facial_expression_recognition/fer/facial_expression_recognition.py", "imports": ["numpy", "cv2", "PIL", "torch", "os"], "module": "modelscope.models.cv.facial_expression_recognition.fer.facial_expression_recognition"}, "('MODELS', 'image-segmentation', 'm2fp')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_human_parsing/m2fp_net.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.image_human_parsing.m2fp_net"}, "('MODELS', 'shop-segmentation', 'shop-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/shop_segmentation/shop_seg_model.py", "imports": ["numpy", "json", "typing", "torch", "PIL", "os"], "module": "modelscope.models.cv.shop_segmentation.shop_seg_model"}, "('MODELS', 'object-detection-3d', 'depe')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/depe_detect.py", "imports": ["numpy", "os", "typing", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.depe_detect"}, "('BBOX_ASSIGNERS', 'default', 'HungarianAssigner3D')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/core/bbox/assigners/hungarian_assigner_3d.py", "imports": ["mmdet", "scipy", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.core.bbox.assigners.hungarian_assigner_3d"}, "('MATCH_COST', 'default', 'BBox3DL1Cost')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/core/bbox/match_costs/match_cost.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.core.bbox.match_costs.match_cost"}, "('BBOX_CODERS', 'default', 'NMSFreeCoder')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/core/bbox/coders/nms_free_coder.py", "imports": ["mmdet", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.core.bbox.coders.nms_free_coder"}, "('DETECTORS', 'default', 'Petr3D')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/detectors/petr3d.py", "imports": ["numpy", "mmdet", "mmdet3d", "torch", "mmcv"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.detectors.petr3d"}, "('HEADS', 'default', 'PETRv2DEDNHead')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/dense_heads/petrv2_dednhead.py", "imports": ["numpy", "mmdet", "math", "mmdet3d", "torch", "copy", "mmcv"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.dense_heads.petrv2_dednhead"}, "('TRANSFORMER', 'default', 'PETRDNTransformer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/utils/petr_transformer.py", "imports": ["copy", "mmcv", "mmdet", "math", "warnings", "typing", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.petr_transformer"}, "('TRANSFORMER_LAYER', 'default', 'PETRTransformerDecoderLayer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/utils/petr_transformer.py", "imports": ["copy", "mmcv", "mmdet", "math", "warnings", "typing", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.petr_transformer"}, "('ATTENTION', 'default', 'PETRMultiheadAttention')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/utils/petr_transformer.py", "imports": ["copy", "mmcv", "mmdet", "math", "warnings", "typing", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.petr_transformer"}, "('TRANSFORMER_LAYER_SEQUENCE', 'default', 'PETRTransformerEncoder')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/utils/petr_transformer.py", "imports": ["copy", "mmcv", "mmdet", "math", "warnings", "typing", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.petr_transformer"}, "('TRANSFORMER_LAYER_SEQUENCE', 'default', 'PETRTransformerDecoder')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/utils/petr_transformer.py", "imports": ["copy", "mmcv", "mmdet", "math", "warnings", "typing", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.petr_transformer"}, "('POSITIONAL_ENCODING', 'default', 'SinePositionalEncoding3D')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/utils/positional_encoding.py", "imports": ["mmcv", "math", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.positional_encoding"}, "('BACKBONES', 'default', 'VoVNet')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/backbones/vovnet.py", "imports": ["mmcv", "collections", "mmdet", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.backbones.vovnet"}, "('NECKS', 'default', 'CPFPN')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/models/necks/cp_fpn.py", "imports": ["mmcv", "mmdet", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.necks.cp_fpn"}, "('PIPELINES', 'default', 'PadMultiViewImage')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/datasets/pipelines/transform_3d.py", "imports": ["PIL", "numpy", "copy", "mmcv", "mmdet", "mmdet3d", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.pipelines.transform_3d"}, "('PIPELINES', 'default', 'NormalizeMultiviewImage')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/datasets/pipelines/transform_3d.py", "imports": ["PIL", "numpy", "copy", "mmcv", "mmdet", "mmdet3d", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.pipelines.transform_3d"}, "('PIPELINES', 'default', 'ResizeCropFlipImage')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/datasets/pipelines/transform_3d.py", "imports": ["PIL", "numpy", "copy", "mmcv", "mmdet", "mmdet3d", "torch"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.pipelines.transform_3d"}, "('PIPELINES', 'default', 'LoadMultiViewImageFromMultiSweepsFiles')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/datasets/pipelines/loading.py", "imports": ["mmcv", "mmdet", "numpy"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.pipelines.loading"}, "('DATASETS', 'default', 'CustomNuScenesDataset')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/object_detection_3d/depe/mmdet3d_plugin/datasets/nuscenes_dataset.py", "imports": ["numpy", "mmdet", "mmdet3d"], "module": "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.nuscenes_dataset"}, "('MODELS', 'video-depth-estimation', 'dro-resnet18-depth-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_depth_estimation/dro_model.py", "imports": ["numpy", "tqdm", "cv2", "torch", "os", "glob"], "module": "modelscope.models.cv.video_depth_estimation.dro_model"}, "('MODELS', 'image-classification', 'EasyRobustModel')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/robust_image_classification/easyrobust_model.py", "imports": ["os", "torch"], "module": "modelscope.models.cv.robust_image_classification.easyrobust_model"}, "('MODELS', 'video-object-detection', 'longshortnet')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_streaming_perception/longshortnet/longshortnet.py", "imports": ["numpy", "time", "tqdm", "json", "argparse", "cv2", "torch", "logging", "os"], "module": "modelscope.models.cv.video_streaming_perception.longshortnet.longshortnet"}, "('MODELS', 'video-deinterlace', 'video-deinterlace')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_deinterlace/UNet_for_video_deinterlace.py", "imports": ["copy", "os", "typing", "torch"], "module": "modelscope.models.cv.video_deinterlace.UNet_for_video_deinterlace"}, "('MODELS', 'panorama-depth-estimation', 's2net-depth-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/s2net_panorama_depth_estimation/s2net_model.py", "imports": ["numpy", "os", "torchvision", "torch"], "module": "modelscope.models.cv.s2net_panorama_depth_estimation.s2net_model"}, "('MODELS', 'text-driven-segmentation', 'text-driven-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/text_driven_segmentation/lseg_model.py", "imports": ["numpy", "json", "typing", "torch", "PIL", "os"], "module": "modelscope.models.cv.text_driven_segmentation.lseg_model"}, "('MODELS', 'image-face-fusion', 'image-face-fusion')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_face_fusion/image_face_fusion.py", "imports": ["numpy", "collections", "torchvision", "typing", "torch", "PIL", "cv2", "os"], "module": "modelscope.models.cv.image_face_fusion.image_face_fusion"}, "('MODELS', 'image-deblurring', 'nafnet')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/image_deblur/nafnet_for_image_deblur.py", "imports": ["os", "typing", "torch"], "module": "modelscope.models.cv.image_deblur.nafnet_for_image_deblur"}, "('MODELS', 'body-3d-keypoints', 'body-3d-keypoints')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/body_3d_keypoints/cannonical_pose/body_3d_pose.py", "imports": ["numpy", "typing", "torch", "logging", "os"], "module": "modelscope.models.cv.body_3d_keypoints.cannonical_pose.body_3d_pose"}, "('MODELS', 'body-3d-keypoints', 'hdformer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/body_3d_keypoints/hdformer/hdformer_detector.py", "imports": ["numpy", "os", "typing", "torch"], "module": "modelscope.models.cv.body_3d_keypoints.hdformer.hdformer_detector"}, "('MODELS', 'video-stabilization', 'video-stabilization')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/video_stabilization/DUTRAFTStabilizer.py", "imports": ["numpy", "math", "typing", "cv2", "torch", "tempfile", "sys", "os"], "module": "modelscope.models.cv.video_stabilization.DUTRAFTStabilizer"}, "('MODELS', 'face-2d-keypoints', 'flc')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/cv/facial_landmark_confidence/flc/facial_landmark_confidence.py", "imports": ["numpy", "cv2", "PIL", "torch", "os"], "module": "modelscope.models.cv.facial_landmark_confidence.flc.facial_landmark_confidence"}, "('MODELS', 'text-to-video-synthesis', 'videocomposer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/videocomposer/videocomposer_model.py", "imports": ["typing", "torch", "einops", "os", "copy", "pynvml", "open_clip"], "module": "modelscope.models.multi_modal.videocomposer.videocomposer_model"}, "('MODELS', 'image-captioning', 'ofa')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["functools", "re", "math", "json", "string", "typing", "torch", "os"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'ocr-recognition', 'ofa')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["functools", "re", "math", "json", "string", "typing", "torch", "os"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'visual-grounding', 'ofa')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["functools", "re", "math", "json", "string", "typing", "torch", "os"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'visual-question-answering', 'ofa')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["functools", "re", "math", "json", "string", "typing", "torch", "os"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'visual-entailment', 'ofa')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["functools", "re", "math", "json", "string", "typing", "torch", "os"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'image-classification', 'ofa')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["functools", "re", "math", "json", "string", "typing", "torch", "os"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'text-summarization', 'ofa')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["functools", "re", "math", "json", "string", "typing", "torch", "os"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'text-classification', 'ofa')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["functools", "re", "math", "json", "string", "typing", "torch", "os"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'auto-speech-recognition', 'ofa')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["functools", "re", "math", "json", "string", "typing", "torch", "os"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'sudoku', 'ofa')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["functools", "re", "math", "json", "string", "typing", "torch", "os"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'text2sql', 'ofa')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/ofa_for_all_tasks.py", "imports": ["functools", "re", "math", "json", "string", "typing", "torch", "os"], "module": "modelscope.models.multi_modal.ofa_for_all_tasks"}, "('MODELS', 'multi-modal-similarity', 'team-multi-modal-similarity')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/team/team_model.py", "imports": ["numpy", "torchvision", "typing", "torch", "PIL", "cv2", "tokenizers"], "module": "modelscope.models.multi_modal.team.team_model"}, "('MODELS', 'text-to-image-synthesis', 'stable-diffusion-xl')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/stable_diffusion/stable_diffusion_xl.py", "imports": ["transformers", "torchvision", "functools", "packaging", "typing", "torch", "os", "diffusers", "random"], "module": "modelscope.models.multi_modal.stable_diffusion.stable_diffusion_xl"}, "('MODELS', 'text-to-image-synthesis', 'stable-diffusion')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/stable_diffusion/stable_diffusion.py", "imports": ["transformers", "functools", "packaging", "typing", "torch", "diffusers", "os"], "module": "modelscope.models.multi_modal.stable_diffusion.stable_diffusion"}, "('MODELS', 'video-multi-modal-embedding', 'video-clip-multi-modal-embedding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mmr/models/clip_for_mm_video_embedding.py", "imports": ["numpy", "decord", "json", "typing", "torch", "PIL", "urllib", "uuid", "tempfile", "os", "random"], "module": "modelscope.models.multi_modal.mmr.models.clip_for_mm_video_embedding"}, "('MODELS', 'multi-modal-embedding', 'clip-multi-modal-embedding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/clip/model.py", "imports": ["numpy", "collections", "json", "typing", "torch", "os"], "module": "modelscope.models.multi_modal.clip.model"}, "('MODELS', 'generative-multi-modal-embedding', 'gemm-generative-multi-modal')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/gemm/gemm_model.py", "imports": ["numpy", "torchvision", "json", "typing", "torch", "PIL", "os"], "module": "modelscope.models.multi_modal.gemm.gemm_model"}, "('MODELS', 'multimodal-dialogue', 'mplug-owl')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mplug_owl/modeling_mplug_owl.py", "imports": ["transformers", "dataclasses", "math", "typing", "torch", "os", "copy", "logging", "random", "io"], "module": "modelscope.models.multi_modal.mplug_owl.modeling_mplug_owl"}, "('MODELS', 'text-to-video-synthesis', 'latent-text-to-video-synthesis')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/video_synthesis/text_to_video_synthesis_model.py", "imports": ["typing", "torch", "einops", "os", "open_clip"], "module": "modelscope.models.multi_modal.video_synthesis.text_to_video_synthesis_model"}, "('MODELS', 'text-video-retrieval', 'prost-clip-text-video-retrieval')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/prost/models/prost_model.py", "imports": ["numpy", "decord", "json", "typing", "torch", "PIL", "urllib", "uuid", "tempfile", "os", "random"], "module": "modelscope.models.multi_modal.prost.models.prost_model"}, "('MODELS', 'efficient-diffusion-tuning', 'efficient-diffusion-tuning')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/efficient_diffusion_tuning/efficient_stable_diffusion.py", "imports": ["transformers", "functools", "typing", "torch", "diffusers", "os", "swift"], "module": "modelscope.models.multi_modal.efficient_diffusion_tuning.efficient_stable_diffusion"}, "('MODELS', 'token-classification', 'mgeo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mgeo/token_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.token_classification"}, "('MODELS', 'part-of-speech', 'mgeo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mgeo/token_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.token_classification"}, "('MODELS', 'word-segmentation', 'mgeo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mgeo/token_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.token_classification"}, "('MODELS', 'backbone', 'mgeo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mgeo/backbone.py", "imports": ["transformers", "dataclasses", "math", "warnings", "torch", "typing", "os", "random"], "module": "modelscope.models.multi_modal.mgeo.backbone"}, "('MODELS', 'text-ranking', 'mgeo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mgeo/text_ranking.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.text_ranking"}, "('MODELS', 'text-classification', 'mgeo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mgeo/text_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.text_classification"}, "('MODELS', 'nli', 'mgeo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mgeo/text_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.text_classification"}, "('MODELS', 'sentiment-classification', 'mgeo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mgeo/text_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.text_classification"}, "('MODELS', 'sentence-similarity', 'mgeo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mgeo/text_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.text_classification"}, "('MODELS', 'zero-shot-classification', 'mgeo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mgeo/text_classification.py", "imports": ["torch"], "module": "modelscope.models.multi_modal.mgeo.text_classification"}, "('MODELS', 'document-vl-embedding', 'vldoc')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/vldoc/model.py", "imports": ["torchvision", "re", "math", "json", "torch", "sys", "copy", "logging", "os"], "module": "modelscope.models.multi_modal.vldoc.model"}, "('MODELS', 'generative-multi-modal-embedding', 'rleg-generative-multi-modal')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/rleg/rleg.py", "imports": ["torchvision", "typing", "torch"], "module": "modelscope.models.multi_modal.rleg.rleg"}, "('MODELS', 'visual-question-answering', 'mplug')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mplug_for_all_tasks.py", "imports": ["os", "typing"], "module": "modelscope.models.multi_modal.mplug_for_all_tasks"}, "('MODELS', 'image-captioning', 'mplug')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mplug_for_all_tasks.py", "imports": ["os", "typing"], "module": "modelscope.models.multi_modal.mplug_for_all_tasks"}, "('MODELS', 'image-text-retrieval', 'mplug')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mplug_for_all_tasks.py", "imports": ["os", "typing"], "module": "modelscope.models.multi_modal.mplug_for_all_tasks"}, "('MODELS', 'video-question-answering', 'hitea')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mplug_for_all_tasks.py", "imports": ["os", "typing"], "module": "modelscope.models.multi_modal.mplug_for_all_tasks"}, "('MODELS', 'video-captioning', 'hitea')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/mplug_for_all_tasks.py", "imports": ["os", "typing"], "module": "modelscope.models.multi_modal.mplug_for_all_tasks"}, "('MODELS', 'image-captioning', 'clip-interrogator')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/clip_interrogator/model.py", "imports": ["numpy", "transformers", "math", "PIL", "hashlib", "requests", "open_clip", "time", "torchvision", "dataclasses", "tqdm", "typing", "torch", "safetensors", "os"], "module": "modelscope.models.multi_modal.clip_interrogator.model"}, "('MODELS', 'text-to-image-synthesis', 'diffusion-text-to-image-synthesis')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/diffusion/model.py", "imports": ["numpy", "json", "typing", "torch", "os"], "module": "modelscope.models.multi_modal.diffusion.model"}, "('MODELS', 'image-to-video', 'image-to-video-model')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/image_to_video/image_to_video_model.py", "imports": ["typing", "torch", "os", "copy", "random"], "module": "modelscope.models.multi_modal.image_to_video.image_to_video_model"}, "('MODELS', 'text-to-image-synthesis', 'multi-stage-diffusion-text-to-image-synthesis')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/multi_stage_diffusion/model.py", "imports": ["numpy", "math", "json", "typing", "torch", "PIL", "os"], "module": "modelscope.models.multi_modal.multi_stage_diffusion.model"}, "('MODELS', 'video-temporal-grounding', 'soonet')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/soonet/model.py", "imports": ["os", "torch"], "module": "modelscope.models.multi_modal.soonet.model"}, "('MODELS', 'text-to-image-synthesis', 'ofa')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/ofa_for_text_to_image_synthesis_model.py", "imports": ["numpy", "torchvision", "json", "typing", "torch", "PIL", "taming", "pkg_resources", "os"], "module": "modelscope.models.multi_modal.ofa_for_text_to_image_synthesis_model"}, "('MODELS', 'video-to-video', 'video-to-video-model')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/models/multi_modal/video_to_video/video_to_video_model.py", "imports": ["typing", "torch", "os", "copy", "random"], "module": "modelscope.models.multi_modal.video_to_video.video_to_video_model"}, "('METRICS', 'default', 'video-super-resolution-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/video_super_resolution_metric/video_super_resolution_metric.py", "imports": ["numpy", "typing"], "module": "modelscope.metrics.video_super_resolution_metric.video_super_resolution_metric"}, "('METRICS', 'default', 'ned')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/ned_metric.py", "imports": ["numpy", "typing"], "module": "modelscope.metrics.ned_metric"}, "('METRICS', 'default', 'image-portrait-enhancement-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/image_portrait_enhancement_metric.py", "imports": ["numpy", "typing", "cv2"], "module": "modelscope.metrics.image_portrait_enhancement_metric"}, "('METRICS', 'default', 'image-quality-assessment-degradation-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/image_quality_assessment_degradation_metric.py", "imports": ["numpy", "collections", "tqdm", "typing", "cv2", "torch", "tempfile", "sys", "os", "scipy"], "module": "modelscope.metrics.image_quality_assessment_degradation_metric"}, "('METRICS', 'default', 'prediction-saving-wrapper')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/prediction_saving_wrapper.py", "imports": ["typing"], "module": "modelscope.metrics.prediction_saving_wrapper"}, "('METRICS', 'default', 'text-ranking-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/text_ranking_metric.py", "imports": ["numpy", "typing"], "module": "modelscope.metrics.text_ranking_metric"}, "('METRICS', 'default', 'video-summarization-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/video_summarization_metric.py", "imports": ["numpy", "typing"], "module": "modelscope.metrics.video_summarization_metric"}, "('METRICS', 'default', 'mAP')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/map_metric.py", "imports": ["numpy", "typing"], "module": "modelscope.metrics.map_metric"}, "('METRICS', 'default', 'ocr-recognition-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/ocr_recognition_metric.py", "imports": ["numpy", "torch", "typing", "edit_distance"], "module": "modelscope.metrics.ocr_recognition_metric"}, "('METRICS', 'default', 'bleu')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/bleu_metric.py", "imports": ["sacrebleu", "itertools", "typing"], "module": "modelscope.metrics.bleu_metric"}, "('METRICS', 'default', 'image-inpainting-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/image_inpainting_metric.py", "imports": ["numpy", "scipy", "typing", "torch"], "module": "modelscope.metrics.image_inpainting_metric"}, "('METRICS', 'default', 'inbatch_recall')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/inbatch_recall_metric.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.metrics.inbatch_recall_metric"}, "('METRICS', 'default', 'token-cls-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/token_classification_metric.py", "imports": ["numpy", "importlib", "typing"], "module": "modelscope.metrics.token_classification_metric"}, "('METRICS', 'default', 'loss-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/loss_metric.py", "imports": ["numpy", "sklearn", "typing"], "module": "modelscope.metrics.loss_metric"}, "('METRICS', 'default', 'seq-cls-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/sequence_classification_metric.py", "imports": ["numpy", "sklearn", "typing"], "module": "modelscope.metrics.sequence_classification_metric"}, "('METRICS', 'default', 'text-gen-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/text_generation_metric.py", "imports": ["nltk", "contextlib", "typing", "sys", "rouge"], "module": "modelscope.metrics.text_generation_metric"}, "('METRICS', 'default', 'image-ins-seg-coco-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/image_instance_segmentation_metric.py", "imports": ["numpy", "collections", "typing", "pycocotools", "tempfile", "os"], "module": "modelscope.metrics.image_instance_segmentation_metric"}, "('METRICS', 'default', 'image-denoise-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/image_denoise_metric.py", "imports": ["numpy", "cv2", "typing", "torch"], "module": "modelscope.metrics.image_denoise_metric"}, "('METRICS', 'default', 'audio-noise-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/audio_noise_metric.py", "imports": ["typing"], "module": "modelscope.metrics.audio_noise_metric"}, "('METRICS', 'default', 'accuracy')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/accuracy_metric.py", "imports": ["numpy", "typing"], "module": "modelscope.metrics.accuracy_metric"}, "('METRICS', 'default', 'translation-evaluation-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/translation_evaluation_metric.py", "imports": ["pandas", "importlib", "typing"], "module": "modelscope.metrics.translation_evaluation_metric"}, "('METRICS', 'default', 'image-color-enhance-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/image_color_enhance_metric.py", "imports": ["numpy", "typing", "cv2"], "module": "modelscope.metrics.image_color_enhance_metric"}, "('METRICS', 'default', 'movie-scene-segmentation-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/movie_scene_segmentation_metric.py", "imports": ["numpy", "typing"], "module": "modelscope.metrics.movie_scene_segmentation_metric"}, "('METRICS', 'default', 'video-stabilization-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/video_stabilization_metric.py", "imports": ["numpy", "tqdm", "typing", "cv2", "tempfile", "sys", "os"], "module": "modelscope.metrics.video_stabilization_metric"}, "('METRICS', 'default', 'ppl')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/ppl_metric.py", "imports": ["numpy", "math", "typing", "torch"], "module": "modelscope.metrics.ppl_metric"}, "('METRICS', 'default', 'image-quality-assessment-mos-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/image_quality_assessment_mos_metric.py", "imports": ["numpy", "tqdm", "typing", "torch", "cv2", "tempfile", "sys", "os", "scipy"], "module": "modelscope.metrics.image_quality_assessment_mos_metric"}, "('METRICS', 'default', 'video-frame-interpolation-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/video_frame_interpolation_metric.py", "imports": ["numpy", "lpips", "math", "typing", "torch"], "module": "modelscope.metrics.video_frame_interpolation_metric"}, "('METRICS', 'default', 'image-colorization-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/image_colorization_metric.py", "imports": ["numpy", "torchvision", "typing", "cv2", "torch", "scipy"], "module": "modelscope.metrics.image_colorization_metric"}, "('METRICS', 'default', 'referring-video-object-segmentation-metric')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/metrics/referring_video_object_segmentation_metric.py", "imports": ["numpy", "tqdm", "typing", "torch", "pycocotools"], "module": "modelscope.metrics.referring_video_object_segmentation_metric"}, "('PIPELINES', 'task-template', 'pipeline-template')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/pipeline_template.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.pipeline_template"}, "('PIPELINES', 'document-segmentation', 'document-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/document_segmentation_pipeline.py", "imports": ["numpy", "re", "typing", "torch", "datasets"], "module": "modelscope.pipelines.nlp.document_segmentation_pipeline"}, "('PIPELINES', 'text-classification', 'user-satisfaction-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/user_satisfaction_estimation_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.nlp.user_satisfaction_estimation_pipeline"}, "('PIPELINES', 'text-summarization', 'text-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/summarization_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.nlp.summarization_pipeline"}, "('PIPELINES', 'document-grounded-dialog-retrieval', 'document-grounded-dialog-retrieval')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/document_grounded_dialog_retrieval_pipeline.py", "imports": ["numpy", "json", "typing", "faiss", "os"], "module": "modelscope.pipelines.nlp.document_grounded_dialog_retrieval_pipeline"}, "('PIPELINES', 'word-segmentation', 'word-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/word_segmentation_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.nlp.word_segmentation_pipeline"}, "('PIPELINES', 'word-segmentation', 'multilingual-word-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/word_segmentation_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.nlp.word_segmentation_pipeline"}, "('PIPELINES', 'word-segmentation', 'word-segmentation-thai')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/word_segmentation_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.nlp.word_segmentation_pipeline"}, "('PIPELINES', 'text-generation', 'gpt-moe-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/distributed_gpt_moe_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.nlp.distributed_gpt_moe_pipeline"}, "('PIPELINES', 'code-translation', 'codegeex-code-translation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/codegeex_code_translation_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.codegeex_code_translation_pipeline"}, "('PIPELINES', 'document-grounded-dialog-rerank', 'document-grounded-dialog-rerank')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/document_grounded_dialog_rerank_pipeline.py", "imports": ["numpy", "time", "pprint", "transformers", "collections", "re", "ujson", "typing", "torch", "sys", "os", "random"], "module": "modelscope.pipelines.nlp.document_grounded_dialog_rerank_pipeline"}, "('PIPELINES', 'text-classification', 'language_identification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/language_identification_pipline.py", "imports": ["numpy", "re", "typing", "os", "tensorflow"], "module": "modelscope.pipelines.nlp.language_identification_pipline"}, "('PIPELINES', 'extractive-summarization', 'extractive-summarization')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/extractive_summarization_pipeline.py", "imports": ["numpy", "re", "typing", "torch", "datasets"], "module": "modelscope.pipelines.nlp.extractive_summarization_pipeline"}, "('PIPELINES', 'zero-shot-classification', 'zero-shot-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/zero_shot_classification_pipeline.py", "imports": ["typing", "scipy", "torch"], "module": "modelscope.pipelines.nlp.zero_shot_classification_pipeline"}, "('PIPELINES', 'table-question-answering', 'table-question-answering-pipeline')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/table_question_answering_pipeline.py", "imports": ["transformers", "json", "typing", "torch", "os"], "module": "modelscope.pipelines.nlp.table_question_answering_pipeline"}, "('PIPELINES', 'named-entity-recognition', 'named-entity-recognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/named_entity_recognition_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.named_entity_recognition_pipeline"}, "('PIPELINES', 'named-entity-recognition', 'named-entity-recognition-thai')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/named_entity_recognition_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.named_entity_recognition_pipeline"}, "('PIPELINES', 'named-entity-recognition', 'named-entity-recognition-viet')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/named_entity_recognition_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.named_entity_recognition_pipeline"}, "('PIPELINES', 'word-alignment', 'word-alignment')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/word_alignment_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.nlp.word_alignment_pipeline"}, "('PIPELINES', 'faq-question-answering', 'faq-question-answering')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/faq_question_answering_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.faq_question_answering_pipeline"}, "('PIPELINES', 'text-classification', 'sentiment-analysis')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_classification_pipeline"}, "('PIPELINES', 'nli', 'nli')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_classification_pipeline"}, "('PIPELINES', 'sentence-similarity', 'sentence-similarity')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_classification_pipeline"}, "('PIPELINES', 'text-classification', 'text-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_classification_pipeline"}, "('PIPELINES', 'text-classification', 'sentiment-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_classification_pipeline"}, "('PIPELINES', 'text-classification', 'sentence-similarity')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_classification_pipeline"}, "('PIPELINES', 'sentiment-classification', 'sentiment-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_classification_pipeline"}, "('PIPELINES', 'text-generation', 'plug-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/distributed_plug_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.nlp.distributed_plug_pipeline"}, "('PIPELINES', 'text-generation', 'text-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'text2text-generation', 'translation_en_to_de')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'text2text-generation', 'translation_en_to_ro')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'text2text-generation', 'translation_en_to_fr')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'text2text-generation', 'text2text-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'chat', 'chatglm6b-text-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'chat', 'chatglm2_6b-text-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'chat', 'qwen-chat')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'text-generation', 'qwen-text-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'text-generation', 'seqgpt')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'text-generation', 'llama2-text-generation-pipeline')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'chat', 'llama2-text-generation-chat-pipeline')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_generation_pipeline.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.pipelines.nlp.text_generation_pipeline"}, "('PIPELINES', 'task-oriented-conversation', 'dialog-modeling')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/dialog_modeling_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.dialog_modeling_pipeline"}, "('PIPELINES', 'task-oriented-conversation', 'dialog-state-tracking')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/dialog_state_tracking_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.dialog_state_tracking_pipeline"}, "('PIPELINES', 'translation', 'automatic-post-editing')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/automatic_post_editing_pipeline.py", "imports": ["html", "numpy", "sentencepiece", "typing", "sacremoses", "jieba", "os", "tensorflow"], "module": "modelscope.pipelines.nlp.automatic_post_editing_pipeline"}, "('PIPELINES', 'task-oriented-conversation', 'dialog-intent-prediction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/dialog_intent_prediction_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.dialog_intent_prediction_pipeline"}, "('PIPELINES', 'text-generation', 'gpt3-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/distributed_gpt3_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.nlp.distributed_gpt3_pipeline"}, "('PIPELINES', 'sentence-embedding', 'sentence-embedding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/sentence_embedding_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.nlp.sentence_embedding_pipeline"}, "('PIPELINES', 'text-generation', 'glm130b-text-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/glm130b_text_generation_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.glm130b_text_generation_pipeline"}, "('PIPELINES', 'fill-mask', 'fill-mask')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/fill_mask_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.nlp.fill_mask_pipeline"}, "('PIPELINES', 'fill-mask', 'fill-mask-ponet')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/fill_mask_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.nlp.fill_mask_pipeline"}, "('PIPELINES', 'document-grounded-dialog-generate', 'document-grounded-dialog-generate')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/document_grounded_dialog_generate_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.document_grounded_dialog_generate_pipeline"}, "('PIPELINES', 'machine-reading-comprehension', 'machine-reading-comprehension-for-ner')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/machine_reading_comprehension_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.nlp.machine_reading_comprehension_pipeline"}, "('PIPELINES', 'text-ranking', 'text-ranking')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_ranking_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.nlp.text_ranking_pipeline"}, "('PIPELINES', 'text-error-correction', 'text-error-correction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/text_error_correction_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.nlp.text_error_correction_pipeline"}, "('PIPELINES', 'text-classification', 'domain-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/fasttext_text_classification_pipeline.py", "imports": ["numpy", "sentencepiece", "typing", "fasttext", "os"], "module": "modelscope.pipelines.nlp.fasttext_text_classification_pipeline"}, "('PIPELINES', 'text-generation', 'polylm-text-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/polylm_text_generation_pipeline.py", "imports": ["os", "typing", "torch"], "module": "modelscope.pipelines.nlp.polylm_text_generation_pipeline"}, "('PIPELINES', 'code-generation', 'codegeex-code-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/codegeex_code_generation_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.nlp.codegeex_code_generation_pipeline"}, "('PIPELINES', 'translation-evaluation', 'translation-evaluation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/translation_evaluation_pipeline.py", "imports": ["numpy", "enum", "typing", "torch", "os"], "module": "modelscope.pipelines.nlp.translation_evaluation_pipeline"}, "('PIPELINES', 'translation', 'interactive-translation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/interactive_translation_pipeline.py", "imports": ["numpy", "typing", "sacremoses", "subword_nmt", "jieba", "os", "tensorflow"], "module": "modelscope.pipelines.nlp.interactive_translation_pipeline"}, "('PIPELINES', 'token-classification', 'token-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/token_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.nlp.token_classification_pipeline"}, "('PIPELINES', 'token-classification', 'part-of-speech')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/token_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.nlp.token_classification_pipeline"}, "('PIPELINES', 'token-classification', 'word-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/token_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.nlp.token_classification_pipeline"}, "('PIPELINES', 'token-classification', 'named-entity-recognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/token_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.nlp.token_classification_pipeline"}, "('PIPELINES', 'part-of-speech', 'part-of-speech')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/token_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.nlp.token_classification_pipeline"}, "('PIPELINES', 'siamese-uie', 'siamese-uie')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/siamese_uie_pipeline.py", "imports": ["time", "math", "tqdm", "json", "typing", "torch", "pathlib", "copy", "logging", "os", "scipy"], "module": "modelscope.pipelines.nlp.siamese_uie_pipeline"}, "('PIPELINES', 'sentence-similarity', 'translation-quality-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/translation_quality_estimation_pipeline.py", "imports": ["transformers", "typing", "torch", "os", "io"], "module": "modelscope.pipelines.nlp.translation_quality_estimation_pipeline"}, "('PIPELINES', 'competency-aware-translation', 'canmt-translation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/canmt_translation_pipeline.py", "imports": ["typing", "os", "sacremoses", "torch"], "module": "modelscope.pipelines.nlp.canmt_translation_pipeline"}, "('PIPELINES', 'table-question-answering', 'conversational-text-to-sql')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/conversational_text_to_sql_pipeline.py", "imports": ["typing", "text2sql_lgesql", "torch"], "module": "modelscope.pipelines.nlp.conversational_text_to_sql_pipeline"}, "('PIPELINES', 'text-summarization', 'mglm-text-summarization')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/mglm_text_summarization_pipeline.py", "imports": ["os", "typing"], "module": "modelscope.pipelines.nlp.mglm_text_summarization_pipeline"}, "('PIPELINES', 'translation', 'csanmt-translation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/translation_pipeline.py", "imports": ["numpy", "typing", "sacremoses", "subword_nmt", "jieba", "os", "tensorflow"], "module": "modelscope.pipelines.nlp.translation_pipeline"}, "('PIPELINES', 'fid-dialogue', 'fid-dialogue')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/fid_dialogue_pipeline.py", "imports": ["re", "typing", "torch"], "module": "modelscope.pipelines.nlp.fid_dialogue_pipeline"}, "('PIPELINES', 'information-extraction', 'relation-extraction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/information_extraction_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.nlp.information_extraction_pipeline"}, "('PIPELINES', 'relation-extraction', 'relation-extraction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/information_extraction_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.nlp.information_extraction_pipeline"}, "('PIPELINES', 'feature-extraction', 'feature-extraction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/nlp/feature_extraction_pipeline.py", "imports": ["os", "typing", "torch"], "module": "modelscope.pipelines.nlp.feature_extraction_pipeline"}, "('PIPELINES', 'protein-structure', 'unifold-protein-structure')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/science/protein_structure_pipeline.py", "imports": ["numpy", "time", "json", "typing", "torch", "os", "unicore"], "module": "modelscope.pipelines.science.protein_structure_pipeline"}, "('PIPELINES', 'speech-language-recognition', 'speech-language-recognition-eres2net')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/language_recognition_eres2net_pipeline.py", "imports": ["numpy", "typing", "torch", "soundfile", "os", "io", "torchaudio"], "module": "modelscope.pipelines.audio.language_recognition_eres2net_pipeline"}, "('PIPELINES', 'auto-speech-recognition', 'asr-inference')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/asr_inference_pipeline.py", "imports": ["yaml", "json", "os", "typing"], "module": "modelscope.pipelines.audio.asr_inference_pipeline"}, "('PIPELINES', 'acoustic-noise-suppression', 'speech_dfsmn_ans_psm_48k_causal')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/ans_dfsmn_pipeline.py", "imports": ["numpy", "collections", "typing", "librosa", "torch", "sys", "soundfile", "os", "io"], "module": "modelscope.pipelines.audio.ans_dfsmn_pipeline"}, "('PIPELINES', 'speaker-verification', 'speaker-verification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/speaker_verification_light_pipeline.py", "imports": ["numpy", "typing", "torch", "soundfile", "os", "io", "torchaudio"], "module": "modelscope.pipelines.audio.speaker_verification_light_pipeline"}, "('PIPELINES', 'speaker-verification', 'speaker-verification-eres2net')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/speaker_verification_eres2net_pipeline.py", "imports": ["soundfile", "io", "typing", "torch"], "module": "modelscope.pipelines.audio.speaker_verification_eres2net_pipeline"}, "('PIPELINES', 'speech-timestamp', 'speech-timestamp-inference')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/timestamp_pipeline.py", "imports": ["yaml", "json", "typing", "funasr", "os"], "module": "modelscope.pipelines.audio.timestamp_pipeline"}, "('PIPELINES', 'text-to-speech', 'sambert-hifigan-tts')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/text_to_speech_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.audio.text_to_speech_pipeline"}, "('PIPELINES', 'speaker-diarization-dialogue-detection', 'speaker-diarization-dialogue-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/speaker_diarization_dialogue_detection_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.audio.speaker_diarization_dialogue_detection_pipeline"}, "('PIPELINES', 'language-score-prediction', 'language-score-prediction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/lm_infer_pipeline.py", "imports": ["os", "typing"], "module": "modelscope.pipelines.audio.lm_infer_pipeline"}, "('PIPELINES', 'speaker-verification', 'sv-inference')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/speaker_verification_pipeline.py", "imports": ["yaml", "shutil", "os", "typing"], "module": "modelscope.pipelines.audio.speaker_verification_pipeline"}, "('PIPELINES', 'speaker-diarization', 'speaker-change-locating')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/speaker_change_locating_pipeline.py", "imports": ["numpy", "typing", "torch", "soundfile", "io", "torchaudio"], "module": "modelscope.pipelines.audio.speaker_change_locating_pipeline"}, "('PIPELINES', 'speaker-diarization', 'segmentation-clustering')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/segmentation_clustering_pipeline.py", "imports": ["numpy", "ast", "typing", "torch", "soundfile", "io", "torchaudio"], "module": "modelscope.pipelines.audio.segmentation_clustering_pipeline"}, "('PIPELINES', 'speaker-diarization-semantic-speaker-turn-detection', 'speaker-diarization-semantic-speaker-turn-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/speaker_diarization_semantic_speaker_turn_detection_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.audio.speaker_diarization_semantic_speaker_turn_detection_pipeline"}, "('PIPELINES', 'acoustic-echo-cancellation', 'speech-dfsmn-aec-psm-16k')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/linear_aec_pipeline.py", "imports": ["numpy", "yaml", "typing", "torch", "importlib", "os", "scipy"], "module": "modelscope.pipelines.audio.linear_aec_pipeline"}, "('PIPELINES', 'keyword-spotting', 'kws-kwsbp')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/kws_kwsbp_pipeline.py", "imports": ["typing", "json", "os"], "module": "modelscope.pipelines.audio.kws_kwsbp_pipeline"}, "('PIPELINES', 'keyword-spotting', 'speech_dfsmn_kws_char_farfield')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/kws_farfield_pipeline.py", "imports": ["wave", "numpy", "typing", "soundfile", "io"], "module": "modelscope.pipelines.audio.kws_farfield_pipeline"}, "('PIPELINES', 'speech-language-recognition', 'speech-language-recognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/language_recognition_pipeline.py", "imports": ["numpy", "typing", "torch", "soundfile", "os", "io", "torchaudio"], "module": "modelscope.pipelines.audio.language_recognition_pipeline"}, "('PIPELINES', 'speech-separation', 'funasr-speech-separation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/speech_separation_pipeline.py", "imports": ["yaml", "json", "typing", "funasr", "os"], "module": "modelscope.pipelines.audio.speech_separation_pipeline"}, "('PIPELINES', 'auto-speech-recognition', 'asr-wenet-inference')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/asr_wenet_inference_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.audio.asr_wenet_inference_pipeline"}, "('PIPELINES', 'punctuation', 'punc-inference')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/punctuation_processing_pipeline.py", "imports": ["yaml", "shutil", "os", "typing"], "module": "modelscope.pipelines.audio.punctuation_processing_pipeline"}, "('PIPELINES', 'acoustic-noise-suppression', 'speech_frcrn_ans_cirm_16k')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/ans_pipeline.py", "imports": ["numpy", "typing", "librosa", "torch", "soundfile", "io"], "module": "modelscope.pipelines.audio.ans_pipeline"}, "('PIPELINES', 'inverse-text-processing', 'itn-inference')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/inverse_text_processing_pipeline.py", "imports": ["yaml", "shutil", "os", "typing"], "module": "modelscope.pipelines.audio.inverse_text_processing_pipeline"}, "('PIPELINES', 'speech-separation', 'speech_mossformer_separation_temporal_8k')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/separation_pipeline.py", "imports": ["numpy", "typing", "torch", "soundfile", "io"], "module": "modelscope.pipelines.audio.separation_pipeline"}, "('PIPELINES', 'speech-separation', 'speech_mossformer2_separation_temporal_8k')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/separation_pipeline.py", "imports": ["numpy", "typing", "torch", "soundfile", "io"], "module": "modelscope.pipelines.audio.separation_pipeline"}, "('PIPELINES', 'speaker-diarization', 'speaker-diarization-inference')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/speaker_diarization_pipeline.py", "imports": ["numpy", "yaml", "json", "typing", "shutil", "os"], "module": "modelscope.pipelines.audio.speaker_diarization_pipeline"}, "('PIPELINES', 'voice-activity-detection', 'vad-inference')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/voice_activity_detection_pipeline.py", "imports": ["yaml", "json", "typing", "funasr", "os"], "module": "modelscope.pipelines.audio.voice_activity_detection_pipeline"}, "('PIPELINES', 'speaker-verification', 'speaker-verification-rdino')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/audio/speaker_verification_rdino_pipeline.py", "imports": ["soundfile", "io", "typing", "torch"], "module": "modelscope.pipelines.audio.speaker_verification_rdino_pipeline"}, "('PIPELINES', 'human-detection', 'resnet18-human-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_detection_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.image_detection_pipeline"}, "('PIPELINES', 'image-object-detection', 'vit-object-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_detection_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.image_detection_pipeline"}, "('PIPELINES', 'image-object-detection', 'abnormal-object-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_detection_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.image_detection_pipeline"}, "('PIPELINES', 'face-recognition', 'manual-face-recognition-frfm')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/face_recognition_onnx_fm_pipeline.py", "imports": ["numpy", "onnxruntime", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.face_recognition_onnx_fm_pipeline"}, "('PIPELINES', 'product-segmentation', 'product-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/product_segmentation_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.product_segmentation_pipeline"}, "('PIPELINES', 'pedestrian-attribute-recognition', 'resnet50_pedestrian-attribute-recognition_image')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/pedestrian_attribute_recognition_pipeline.py", "imports": ["numpy", "torchvision", "json", "typing", "torch", "cv2", "PIL", "os"], "module": "modelscope.pipelines.cv.pedestrian_attribute_recognition_pipeline"}, "('PIPELINES', 'image-quality-assessment-degradation', 'image-quality-assessment-degradation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_quality_assessment_degradation_pipeline.py", "imports": ["numpy", "torchvision", "math", "typing", "torch", "cv2", "tempfile"], "module": "modelscope.pipelines.cv.image_quality_assessment_degradation_pipeline"}, "('PIPELINES', 'video-frame-interpolation', 'video-frame-interpolation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_frame_interpolation_pipeline.py", "imports": ["numpy", "torchvision", "subprocess", "math", "typing", "cv2", "torch", "tempfile", "os", "glob"], "module": "modelscope.pipelines.cv.video_frame_interpolation_pipeline"}, "('PIPELINES', 'image-quality-assessment-mos', 'image-quality-assessment-man')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_quality_assessment_man_pipeline.py", "imports": ["numpy", "torchvision", "math", "typing", "torch", "cv2", "tempfile"], "module": "modelscope.pipelines.cv.image_quality_assessment_man_pipeline"}, "('PIPELINES', 'face-quality-assessment', 'manual-face-quality-assessment-fqa')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/face_quality_assessment_pipeline.py", "imports": ["numpy", "onnxruntime", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.face_quality_assessment_pipeline"}, "('PIPELINES', 'object-detection-3d', 'object-detection-3d-depe')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/object_detection_3d_pipeline.py", "imports": ["numpy", "typing", "torch", "PIL", "cv2", "tempfile", "os"], "module": "modelscope.pipelines.cv.object_detection_3d_pipeline"}, "('PIPELINES', 'video-instance-segmentation', 'video-instance-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_instance_segmentation_pipeline.py", "imports": ["numpy", "tqdm", "typing", "torch", "cv2", "mmcv", "os"], "module": "modelscope.pipelines.cv.video_instance_segmentation_pipeline"}, "('PIPELINES', 'face-attribute-recognition', 'resnet34-face-attribute-recognition-fairface')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/face_attribute_recognition_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.face_attribute_recognition_pipeline"}, "('PIPELINES', 'video-depth-estimation', 'video-depth-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_depth_estimation_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.video_depth_estimation_pipeline"}, "('PIPELINES', 'image-skychange', 'image-skychange')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_skychange_pipeline.py", "imports": ["numpy", "time", "typing", "cv2", "PIL", "pdb"], "module": "modelscope.pipelines.cv.image_skychange_pipeline"}, "('PIPELINES', 'video-embedding', 'hicossl-s3dg-video_embedding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/hicossl_video_embedding_pipeline.py", "imports": ["math", "os", "typing", "torch"], "module": "modelscope.pipelines.cv.hicossl_video_embedding_pipeline"}, "('PIPELINES', 'video-single-object-tracking', 'procontext-vitb-video-single-object-tracking')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_single_object_tracking_pipeline.py", "imports": ["os", "typing", "cv2"], "module": "modelscope.pipelines.cv.video_single_object_tracking_pipeline"}, "('PIPELINES', 'video-single-object-tracking', 'ostrack-vitb-video-single-object-tracking')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_single_object_tracking_pipeline.py", "imports": ["os", "typing", "cv2"], "module": "modelscope.pipelines.cv.video_single_object_tracking_pipeline"}, "('PIPELINES', 'card-detection', 'resnet-card-detection-scrfd34gkps')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/card_detection_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.card_detection_pipeline"}, "('PIPELINES', 'vision-efficient-tuning', 'vision-efficient-tuning')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/vision_efficient_tuning_pipeline.py", "imports": ["numpy", "torchvision", "typing", "torch"], "module": "modelscope.pipelines.cv.vision_efficient_tuning_pipeline"}, "('PIPELINES', 'video-object-segmentation', 'video-object-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_object_segmentation_pipeline.py", "imports": ["numpy", "torchvision", "typing", "torch", "PIL", "os"], "module": "modelscope.pipelines.cv.video_object_segmentation_pipeline"}, "('PIPELINES', 'hand-static', 'hand-static')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/hand_static_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.hand_static_pipeline"}, "('PIPELINES', 'image-colorization', 'ddcolor-image-colorization')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/ddcolor_image_colorization_pipeline.py", "imports": ["numpy", "torchvision", "typing", "cv2", "torch"], "module": "modelscope.pipelines.cv.ddcolor_image_colorization_pipeline"}, "('PIPELINES', 'face-recognition', 'ir101-face-recognition-cfglint')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/face_recognition_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.face_recognition_pipeline"}, "('PIPELINES', 'image-style-transfer', 'AAMS-style-transfer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_style_transfer_pipeline.py", "imports": ["numpy", "os", "typing", "cv2"], "module": "modelscope.pipelines.cv.image_style_transfer_pipeline"}, "('PIPELINES', 'body-3d-keypoints', 'canonical_body-3d-keypoints_video')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/body_3d_keypoints_pipeline.py", "imports": ["numpy", "mpl_toolkits", "typing", "cv2", "torch", "tempfile", "datetime", "os", "matplotlib"], "module": "modelscope.pipelines.cv.body_3d_keypoints_pipeline"}, "('PIPELINES', 'image-segmentation', 'm2fp-image-human-parsing')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_human_parsing_pipeline.py", "imports": ["numpy", "torchvision", "typing", "torch"], "module": "modelscope.pipelines.cv.image_human_parsing_pipeline"}, "('PIPELINES', 'image-portrait-stylization', 'unet-person-image-cartoon')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_cartoon_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "tensorflow"], "module": "modelscope.pipelines.cv.image_cartoon_pipeline"}, "('PIPELINES', 'image-depth-estimation', 'image-bts-depth-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_bts_depth_estimation_pipeline.py", "imports": ["numpy", "typing", "cv2", "torch", "albumentations"], "module": "modelscope.pipelines.cv.image_bts_depth_estimation_pipeline"}, "('PIPELINES', 'image-super-resolution', 'mobile-image-super-resolution')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/mobile_image_super_resolution_pipeline.py", "imports": ["numpy", "torchvision", "typing", "torch", "skimage"], "module": "modelscope.pipelines.cv.mobile_image_super_resolution_pipeline"}, "('PIPELINES', 'video-object-detection', 'cspnet_realtime-video-object-detection_streamyolo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/realtime_video_object_detection_pipeline.py", "imports": ["numpy", "torchvision", "json", "typing", "torch", "cv2", "PIL", "os"], "module": "modelscope.pipelines.cv.realtime_video_object_detection_pipeline"}, "('PIPELINES', 'general-recognition', 'resnet101-general-recognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/general_recognition_pipeline.py", "imports": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "os"], "module": "modelscope.pipelines.cv.general_recognition_pipeline"}, "('PIPELINES', 'image-try-on', 'image-try-on')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_try_on_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.cv.image_try_on_pipeline"}, "('PIPELINES', 'image-classification', 'resnet50-image-classification-cc')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/content_check_pipeline.py", "imports": ["numpy", "torchvision", "typing", "torch", "PIL", "cv2", "os"], "module": "modelscope.pipelines.cv.content_check_pipeline"}, "('PIPELINES', 'nerf-recon-acc', 'nerf-recon-acc')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/nerf_recon_acc_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.nerf_recon_acc_pipeline"}, "('PIPELINES', 'ocr-recognition', 'convnextTiny-ocr-recognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/ocr_recognition_pipeline.py", "imports": [], "module": "modelscope.pipelines.cv.ocr_recognition_pipeline"}, "('PIPELINES', 'action-detection', 'ResNetC3D-action-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/action_detection_pipeline.py", "imports": ["math", "os", "typing"], "module": "modelscope.pipelines.cv.action_detection_pipeline"}, "('PIPELINES', 'animal-recognition', 'resnet101-animal-recognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/animal_recognition_pipeline.py", "imports": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "os"], "module": "modelscope.pipelines.cv.animal_recognition_pipeline"}, "('PIPELINES', 'face-liveness', 'manual-face-liveness-flxc')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/face_liveness_xc_pipeline.py", "imports": ["numpy", "onnxruntime", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.face_liveness_xc_pipeline"}, "('PIPELINES', 'human-reconstruction', 'human-reconstruction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/human_reconstruction_pipeline.py", "imports": ["numpy", "typing", "shutil", "torch", "os", "trimesh"], "module": "modelscope.pipelines.cv.human_reconstruction_pipeline"}, "('PIPELINES', 'semantic-segmentation', 'u2net-salient-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_salient_detection_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.image_salient_detection_pipeline"}, "('PIPELINES', 'semantic-segmentation', 'res2net-salient-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_salient_detection_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.image_salient_detection_pipeline"}, "('PIPELINES', 'semantic-segmentation', 'res2net-camouflaged-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_salient_detection_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.image_salient_detection_pipeline"}, "('PIPELINES', 'table-recognition', 'dla34-table-recognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/table_recognition_pipeline.py", "imports": ["numpy", "math", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.table_recognition_pipeline"}, "('PIPELINES', 'semantic-segmentation', 'ddpm-image-semantic-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/ddpm_semantic_segmentation_pipeline.py", "imports": ["torchvision", "typing", "torch"], "module": "modelscope.pipelines.cv.ddpm_semantic_segmentation_pipeline"}, "('PIPELINES', 'video-text-retrieval', 'vop-video-text-retrieval')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/vop_retrieval_pipeline.py", "imports": ["numpy", "gzip", "collections", "math", "pickle", "tqdm", "typing", "torch", "os", "random"], "module": "modelscope.pipelines.cv.vop_retrieval_pipeline"}, "('PIPELINES', 'skin-retouching', 'unet-skin-retouching')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/skin_retouching_pipeline.py", "imports": ["numpy", "torchvision", "typing", "cv2", "PIL", "torch", "os", "tensorflow"], "module": "modelscope.pipelines.cv.skin_retouching_pipeline"}, "('PIPELINES', 'image-portrait-enhancement', 'gpen-image-portrait-enhancement')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_portrait_enhancement_pipeline.py", "imports": ["numpy", "math", "typing", "torch", "PIL", "cv2", "scipy"], "module": "modelscope.pipelines.cv.image_portrait_enhancement_pipeline"}, "('PIPELINES', 'image-to-image-translation', 'image-to-image-translation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_to_image_translation_pipeline.py", "imports": ["numpy", "torchvision", "typing", "cv2", "torch", "PIL", "sys", "os", "io"], "module": "modelscope.pipelines.cv.image_to_image_translation_pipeline"}, "('PIPELINES', 'face-human-hand-detection', 'face-human-hand-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/face_human_hand_detection_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.face_human_hand_detection_pipeline"}, "('PIPELINES', 'image-inpainting', 'image-inpainting-sdv2')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_inpainting_sdv2_pipeline.py", "imports": ["numpy", "math", "typing", "cv2", "torch", "tempfile", "sys", "diffusers", "os"], "module": "modelscope.pipelines.cv.image_inpainting_sdv2_pipeline"}, "('PIPELINES', 'portrait-matting', 'unet-image-matting')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_matting_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "tensorflow"], "module": "modelscope.pipelines.cv.image_matting_pipeline"}, "('PIPELINES', 'universal-matting', 'unet-universal-matting')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_matting_pipeline.py", "imports": ["numpy", "typing", "cv2", "os", "tensorflow"], "module": "modelscope.pipelines.cv.image_matting_pipeline"}, "('PIPELINES', 'image-depth-estimation', 'image-depth-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_depth_estimation_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch"], "module": "modelscope.pipelines.cv.image_depth_estimation_pipeline"}, "('PIPELINES', 'image-classification', 'tinynas-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/tinynas_classification_pipeline.py", "imports": ["torchvision", "math", "typing", "torch", "os"], "module": "modelscope.pipelines.cv.tinynas_classification_pipeline"}, "('PIPELINES', 'image-super-resolution-pasd', 'image-super-resolution-pasd')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_super_resolution_pasd_pipeline.py", "imports": ["numpy", "transformers", "torchvision", "typing", "torch", "PIL", "tempfile", "diffusers", "os"], "module": "modelscope.pipelines.cv.image_super_resolution_pasd_pipeline"}, "('PIPELINES', 'pointcloud-sceneflow-estimation', 'pointcloud-sceneflow-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/pointcloud_sceneflow_estimation_pipeline.py", "imports": ["numpy", "plyfile", "typing", "torch"], "module": "modelscope.pipelines.cv.pointcloud_sceneflow_estimation_pipeline"}, "('PIPELINES', 'image-paintbyexample', 'stablediffusion-paintbyexample')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_paintbyexample_pipeline.py", "imports": ["numpy", "torchvision", "typing", "torch", "PIL", "cv2", "einops"], "module": "modelscope.pipelines.cv.image_paintbyexample_pipeline"}, "('PIPELINES', 'face-recognition', 'manual-face-recognition-frir')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/face_recognition_onnx_ir_pipeline.py", "imports": ["numpy", "onnxruntime", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.face_recognition_onnx_ir_pipeline"}, "('PIPELINES', 'image-classification', 'image-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'image-classification', 'vit-base_image-classification_ImageNet-labels')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'image-classification', 'vit-base_image-classification_Dailylife-labels')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'image-classification', 'nextvit-small_image-classification_Dailylife-labels')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'image-classification', 'convnext-base_image-classification_garbage')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'image-classification', 'common-image-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'image-classification', 'easyrobust-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'image-classification', 'bnext-small_image-classification_ImageNet-labels')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_classification_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.cv.image_classification_pipeline"}, "('PIPELINES', 'image-to-image-generation', 'image-to-image-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_to_image_generate_pipeline.py", "imports": ["numpy", "torchvision", "typing", "torch", "PIL", "cv2", "os"], "module": "modelscope.pipelines.cv.image_to_image_generate_pipeline"}, "('PIPELINES', 'face-reconstruction', 'resnet50-face-reconstruction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/face_reconstruction_pipeline.py", "imports": ["numpy", "shutil", "cv2", "PIL", "torch", "tensorflow", "face_alignment", "typing", "scipy", "os", "io"], "module": "modelscope.pipelines.cv.face_reconstruction_pipeline"}, "('PIPELINES', 'movie-scene-segmentation', 'resnet50-bert-movie-scene-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/movie_scene_segmentation_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.cv.movie_scene_segmentation_pipeline"}, "('PIPELINES', 'video-text-retrieval', 'vop-video-text-retrieval-se')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/vop_retrieval_se_pipeline.py", "imports": ["numpy", "gzip", "typing", "torch", "os"], "module": "modelscope.pipelines.cv.vop_retrieval_se_pipeline"}, "('PIPELINES', 'indoor-layout-estimation', 'indoor-layout-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/indoor_layout_estimation_pipeline.py", "imports": ["numpy", "typing", "cv2"], "module": "modelscope.pipelines.cv.indoor_layout_estimation_pipeline"}, "('PIPELINES', 'image-driving-perception', 'yolopv2_image-driving-percetion_bdd100k')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_driving_perception_pipeline.py", "imports": ["numpy", "os", "typing", "cv2"], "module": "modelscope.pipelines.cv.image_driving_perception_pipeline"}, "('PIPELINES', 'panorama-depth-estimation', 'panorama-depth-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/panorama_depth_estimation_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch"], "module": "modelscope.pipelines.cv.panorama_depth_estimation_pipeline"}, "('PIPELINES', 'video-inpainting', 'video-inpainting')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_inpainting_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.video_inpainting_pipeline"}, "('PIPELINES', 'text-driven-segmentation', 'text-driven-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/text_driven_segmentation_pipleline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.text_driven_segmentation_pipleline"}, "('PIPELINES', 'image-colorization', 'unet-image-colorization')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_colorization_pipeline.py", "imports": ["numpy", "torchvision", "typing", "torch", "PIL", "cv2"], "module": "modelscope.pipelines.cv.image_colorization_pipeline"}, "('PIPELINES', 'nerf-recon-4k', 'nerf-recon-4k')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/nerf_recon_4k_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.nerf_recon_4k_pipeline"}, "('PIPELINES', 'action-recognition', 'TAdaConv_action-recognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/action_recognition_pipeline.py", "imports": ["math", "os", "typing", "torch"], "module": "modelscope.pipelines.cv.action_recognition_pipeline"}, "('PIPELINES', 'action-recognition', 'patchshift-action-recognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/action_recognition_pipeline.py", "imports": ["math", "os", "typing", "torch"], "module": "modelscope.pipelines.cv.action_recognition_pipeline"}, "('PIPELINES', 'image-segmentation', 'cascade-mask-rcnn-swin-image-instance-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_instance_segmentation_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.image_instance_segmentation_pipeline"}, "('PIPELINES', 'language-guided-video-summarization', 'clip-it-video-summarization')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/language_guided_video_summarization_pipeline.py", "imports": ["numpy", "shutil", "cv2", "torch", "typing", "PIL", "tempfile", "os", "clip", "random"], "module": "modelscope.pipelines.cv.language_guided_video_summarization_pipeline"}, "('PIPELINES', 'nerf-recon-vq-compression', 'nerf-recon-vq-compression')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/nerf_recon_vq_compression_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.nerf_recon_vq_compression_pipeline"}, "('PIPELINES', 'shop-segmentation', 'shop-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/shop_segmentation_pipleline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.shop_segmentation_pipleline"}, "('PIPELINES', 'facial-expression-recognition', 'vgg19-facial-expression-recognition-fer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/facial_expression_recognition_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.facial_expression_recognition_pipeline"}, "('PIPELINES', 'image-deblurring', 'nafnet-image-deblur')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_deblur_pipeline.py", "imports": ["torchvision", "typing", "torch"], "module": "modelscope.pipelines.cv.image_deblur_pipeline"}, "('PIPELINES', 'video-panoptic-segmentation', 'video-panoptic-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_panoptic_segmentation_pipeline.py", "imports": ["numpy", "tqdm", "typing", "torch", "cv2", "mmcv", "os"], "module": "modelscope.pipelines.cv.video_panoptic_segmentation_pipeline"}, "('PIPELINES', 'body-2d-keypoints', 'hrnetv2w32_body-2d-keypoints_image')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/body_2d_keypoints_pipeline.py", "imports": ["numpy", "torchvision", "json", "typing", "torch", "cv2", "PIL", "os"], "module": "modelscope.pipelines.cv.body_2d_keypoints_pipeline"}, "('PIPELINES', 'video-stabilization', 'video-stabilization')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_stabilization_pipeline.py", "imports": ["numpy", "subprocess", "math", "typing", "cv2", "torch", "tempfile", "os", "glob"], "module": "modelscope.pipelines.cv.video_stabilization_pipeline"}, "('PIPELINES', 'open-vocabulary-detection', 'open-vocabulary-detection-vild')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_open_vocabulary_detection_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.image_open_vocabulary_detection_pipeline"}, "('PIPELINES', 'domain-specific-object-detection', 'tinynas-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/tinynas_detection_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.tinynas_detection_pipeline"}, "('PIPELINES', 'image-object-detection', 'tinynas-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/tinynas_detection_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.tinynas_detection_pipeline"}, "('PIPELINES', 'video-super-resolution', 'realbasicvsr-video-super-resolution')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_super_resolution_pipeline.py", "imports": ["numpy", "torchvision", "subprocess", "math", "typing", "cv2", "torch", "tempfile", "os"], "module": "modelscope.pipelines.cv.video_super_resolution_pipeline"}, "('PIPELINES', 'face-recognition', 'ir50-face-recognition-arcface')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/arc_face_recognition_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.arc_face_recognition_pipeline"}, "('PIPELINES', 'image-segmentation', 'maskdino-swin-image-instance-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/maskdino_instance_segmentation_pipeline.py", "imports": ["torchvision", "typing", "torch"], "module": "modelscope.pipelines.cv.maskdino_instance_segmentation_pipeline"}, "('PIPELINES', 'video-colorization', 'video-colorization')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_colorization_pipeline.py", "imports": ["numpy", "torchvision", "subprocess", "typing", "cv2", "PIL", "torch", "tempfile", "os"], "module": "modelscope.pipelines.cv.video_colorization_pipeline"}, "('PIPELINES', 'face-emotion', 'face-emotion')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/face_emotion_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.face_emotion_pipeline"}, "('PIPELINES', 'image-object-detection', 'tbs-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/tbs_detection_pipeline.py", "imports": ["numpy", "typing", "torch", "cv2", "PIL", "colorsys", "os"], "module": "modelscope.pipelines.cv.tbs_detection_pipeline"}, "('PIPELINES', 'image-demoireing', 'uhdm-image-demoireing')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_restoration_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.image_restoration_pipeline"}, "('PIPELINES', 'image-debanding', 'rrdb-image-debanding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_debanding_pipeline.py", "imports": ["torchvision", "typing", "torch"], "module": "modelscope.pipelines.cv.image_debanding_pipeline"}, "('PIPELINES', 'image-editing', 'masactrl-image-editing')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_editing_pipeline.py", "imports": ["numpy", "torchvision", "tqdm", "typing", "torch", "PIL", "diffusers", "os"], "module": "modelscope.pipelines.cv.image_editing_pipeline"}, "('PIPELINES', 'image-super-resolution', 'rrdb-image-super-resolution')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_super_resolution_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch"], "module": "modelscope.pipelines.cv.image_super_resolution_pipeline"}, "('PIPELINES', 'video-embedding', 'cmdssl-r2p1d_video_embedding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/cmdssl_video_embedding_pipeline.py", "imports": ["numpy", "torchvision", "decord", "typing", "torch", "PIL", "os"], "module": "modelscope.pipelines.cv.cmdssl_video_embedding_pipeline"}, "('PIPELINES', 'video-multi-object-tracking', 'video-multi-object-tracking')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_multi_object_tracking_pipeline.py", "imports": ["os", "typing", "torch"], "module": "modelscope.pipelines.cv.video_multi_object_tracking_pipeline"}, "('PIPELINES', 'image-face-fusion', 'image-face-fusion')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_face_fusion_pipeline.py", "imports": ["numpy", "typing"], "module": "modelscope.pipelines.cv.image_face_fusion_pipeline"}, "('PIPELINES', 'face-liveness', 'manual-face-liveness-flir')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/face_liveness_ir_pipeline.py", "imports": ["numpy", "onnxruntime", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.face_liveness_ir_pipeline"}, "('PIPELINES', 'image-denoising', 'nafnet-image-denoise')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_denoise_pipeline.py", "imports": ["torchvision", "typing", "torch"], "module": "modelscope.pipelines.cv.image_denoise_pipeline"}, "('PIPELINES', 'panorama-depth-estimation', 'panorama-depth-estimation-s2net')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/panorama_depth_estimation_s2net_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch"], "module": "modelscope.pipelines.cv.panorama_depth_estimation_s2net_pipeline"}, "('PIPELINES', 'image-reid-person', 'passvitb-image-reid-person')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_reid_person_pipeline.py", "imports": ["torchvision", "math", "typing", "torch", "PIL", "os"], "module": "modelscope.pipelines.cv.image_reid_person_pipeline"}, "('PIPELINES', 'image-body-reshaping', 'flow-based-body-reshaping')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_body_reshaping_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.image_body_reshaping_pipeline"}, "('PIPELINES', 'face-detection', 'manual-face-detection-mtcnn')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/mtcnn_face_detection_pipeline.py", "imports": ["os", "typing", "torch"], "module": "modelscope.pipelines.cv.mtcnn_face_detection_pipeline"}, "('PIPELINES', 'face-detection', 'resnet-face-detection-scrfd10gkps')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/face_detection_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.face_detection_pipeline"}, "('PIPELINES', 'ocr-detection', 'resnet18-ocr-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/ocr_detection_pipeline.py", "imports": ["numpy", "math", "typing", "torch", "cv2", "os", "tf_slim", "tensorflow"], "module": "modelscope.pipelines.cv.ocr_detection_pipeline"}, "('PIPELINES', 'video-deinterlace', 'video-deinterlace')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_deinterlace_pipeline.py", "imports": ["numpy", "torchvision", "subprocess", "math", "typing", "cv2", "torch", "tempfile", "os"], "module": "modelscope.pipelines.cv.video_deinterlace_pipeline"}, "('PIPELINES', 'product-retrieval-embedding', 'resnet50-product-retrieval-embedding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/product_retrieval_embedding_pipeline.py", "imports": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "os"], "module": "modelscope.pipelines.cv.product_retrieval_embedding_pipeline"}, "('PIPELINES', 'license-plate-detection', 'resnet18-license-plate-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/license_plate_detection_pipeline.py", "imports": ["numpy", "math", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.license_plate_detection_pipeline"}, "('PIPELINES', 'video-category', 'video-category')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_category_pipeline.py", "imports": ["numpy", "torchvision", "decord", "json", "typing", "torch", "PIL", "os"], "module": "modelscope.pipelines.cv.video_category_pipeline"}, "('PIPELINES', 'image-color-enhancement', 'adaint-image-color-enhance')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_color_enhance_pipeline.py", "imports": ["torchvision", "typing", "torch"], "module": "modelscope.pipelines.cv.image_color_enhance_pipeline"}, "('PIPELINES', 'image-color-enhancement', 'deeplpf-image-color-enhance')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_color_enhance_pipeline.py", "imports": ["torchvision", "typing", "torch"], "module": "modelscope.pipelines.cv.image_color_enhance_pipeline"}, "('PIPELINES', 'image-color-enhancement', 'csrnet-image-color-enhance')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_color_enhance_pipeline.py", "imports": ["torchvision", "typing", "torch"], "module": "modelscope.pipelines.cv.image_color_enhance_pipeline"}, "('PIPELINES', 'referring-video-object-segmentation', 'referring-video-object-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/referring_video_object_segmentation_pipeline.py", "imports": ["numpy", "torchvision", "tqdm", "typing", "torch", "moviepy", "PIL", "tempfile", "einops"], "module": "modelscope.pipelines.cv.referring_video_object_segmentation_pipeline"}, "('PIPELINES', 'virtual-try-on', 'virtual-try-on')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/virtual_try_on_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.virtual_try_on_pipeline"}, "('PIPELINES', 'crowd-counting', 'hrnet-crowd-counting')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/crowd_counting_pipeline.py", "imports": ["numpy", "torchvision", "math", "typing", "torch", "PIL"], "module": "modelscope.pipelines.cv.crowd_counting_pipeline"}, "('PIPELINES', 'image-object-detection', 'vidt')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/vidt_pipeline.py", "imports": ["torchvision", "typing", "torch"], "module": "modelscope.pipelines.cv.vidt_pipeline"}, "('PIPELINES', 'image-segmentation', 'image-panoptic-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_panoptic_segmentation_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch"], "module": "modelscope.pipelines.cv.image_panoptic_segmentation_pipeline"}, "('PIPELINES', 'image-multi-view-depth-estimation', 'image-multi-view-depth-estimation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_mvs_depth_estimation_pipeline.py", "imports": ["shutil", "tempfile", "os", "typing"], "module": "modelscope.pipelines.cv.image_mvs_depth_estimation_pipeline"}, "('PIPELINES', 'bad-image-detecting', 'bad-image-detecting')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/bad_image_detecting_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.cv.bad_image_detecting_pipeline"}, "('PIPELINES', 'text-to-360panorama-image', 'text-to-360panorama-image')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/text_to_360panorama_image_pipeline.py", "imports": ["numpy", "typing", "torch", "basicsr", "PIL", "diffusers", "random", "realesrgan"], "module": "modelscope.pipelines.cv.text_to_360panorama_image_pipeline"}, "('PIPELINES', 'image-matching', 'image-matching')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_matching_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch"], "module": "modelscope.pipelines.cv.image_matching_pipeline"}, "('PIPELINES', 'image-classification', 'image-structured-model-probing')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_structured_model_probing_pipeline.py", "imports": ["numpy", "torchvision", "math", "typing", "torch", "mmcv", "os"], "module": "modelscope.pipelines.cv.image_structured_model_probing_pipeline"}, "('PIPELINES', 'face-recognition', 'ir-face-recognition-rts')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/face_recognition_ood_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.face_recognition_ood_pipeline"}, "('PIPELINES', 'human-image-generation', 'human-image-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/human_image_generation_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.cv.human_image_generation_pipeline"}, "('PIPELINES', 'video-summarization', 'googlenet_pgl_video_summarization')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_summarization_pipeline.py", "imports": ["numpy", "tqdm", "typing", "cv2", "torch", "os"], "module": "modelscope.pipelines.cv.video_summarization_pipeline"}, "('PIPELINES', 'video-human-matting', 'video-human-matting')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/video_human_matting_pipeline.py", "imports": ["numpy", "moviepy", "cv2", "typing", "torch", "os"], "module": "modelscope.pipelines.cv.video_human_matting_pipeline"}, "('PIPELINES', 'image-segmentation', 'fast-instance-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/fast_instance_segmentation_pipeline.py", "imports": ["numpy", "torchvision", "typing", "torch"], "module": "modelscope.pipelines.cv.fast_instance_segmentation_pipeline"}, "('PIPELINES', 'image-fewshot-detection', 'image-fewshot-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_defrcn_fewshot_pipeline.py", "imports": ["numpy", "os", "typing", "torch"], "module": "modelscope.pipelines.cv.image_defrcn_fewshot_pipeline"}, "('PIPELINES', 'face-2d-keypoints', 'manual-facial-landmark-confidence-flcm')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/facial_landmark_confidence_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.facial_landmark_confidence_pipeline"}, "('PIPELINES', 'lineless-table-recognition', 'lore-lineless-table-recognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/lineless_table_recognition_pipeline.py", "imports": ["numpy", "math", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.lineless_table_recognition_pipeline"}, "('PIPELINES', 'surface-recon-common', 'surface-recon-common')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/surface_recon_common_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.cv.surface_recon_common_pipeline"}, "('PIPELINES', 'image-quality-assessment-mos', 'image-quality-assessment-mos')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_quality_assessment_mos_pipeline.py", "imports": ["numpy", "torchvision", "math", "typing", "torch", "cv2", "tempfile"], "module": "modelscope.pipelines.cv.image_quality_assessment_mos_pipeline"}, "('PIPELINES', 'face-detection', 'manual-face-detection-ulfd')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/ulfd_face_detection_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.ulfd_face_detection_pipeline"}, "('PIPELINES', 'controllable-image-generation', 'controllable-image-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/controllable_image_generation_pipeline.py", "imports": ["numpy", "subprocess", "math", "typing", "cv2", "torch", "tempfile", "os", "glob"], "module": "modelscope.pipelines.cv.controllable_image_generation_pipeline"}, "('PIPELINES', 'image-segmentation', 'image-semantic-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_semantic_segmentation_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch"], "module": "modelscope.pipelines.cv.image_semantic_segmentation_pipeline"}, "('PIPELINES', 'image-inpainting', 'fft-inpainting')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/image_inpainting_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch"], "module": "modelscope.pipelines.cv.image_inpainting_pipeline"}, "('PIPELINES', 'image-segmentation', 'vision-middleware-multi-task')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/vision_middleware_pipeline.py", "imports": ["numpy", "torchvision", "math", "typing", "torch", "mmcv", "os"], "module": "modelscope.pipelines.cv.vision_middleware_pipeline"}, "('PIPELINES', 'face-detection', 'resnet50-face-detection-retinaface')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/retina_face_detection_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.retina_face_detection_pipeline"}, "('PIPELINES', 'live-category', 'live-category')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/live_category_pipeline.py", "imports": ["numpy", "torchvision", "decord", "typing", "torch", "PIL", "os"], "module": "modelscope.pipelines.cv.live_category_pipeline"}, "('PIPELINES', 'face-image-generation', 'gan-face-image-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/face_image_generation_pipeline.py", "imports": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "module": "modelscope.pipelines.cv.face_image_generation_pipeline"}, "('PIPELINES', 'face-recognition', 'resnet-face-recognition-facemask')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/mask_face_recognition_pipeline.py", "imports": ["numpy", "collections", "typing", "torch", "PIL", "cv2", "os"], "module": "modelscope.pipelines.cv.mask_face_recognition_pipeline"}, "('PIPELINES', 'motion-generation', 'mdm-motion-generation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/motion_generation_pipeline.py", "imports": ["numpy", "typing", "torch", "tempfile", "os"], "module": "modelscope.pipelines.cv.motion_generation_pipeline"}, "('PIPELINES', 'face-detection', 'resnet101-face-detection-cvpr22papermogface')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/cv/mog_face_detection_pipeline.py", "imports": ["numpy", "os", "typing"], "module": "modelscope.pipelines.cv.mog_face_detection_pipeline"}, "('PIPELINES', 'text-video-retrieval', 'prost-text-video-retrieval')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/prost_text_video_retrieval_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.multi_modal.prost_text_video_retrieval_pipeline"}, "('PIPELINES', 'auto-speech-recognition', 'ofa-asr')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/asr_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.multi_modal.asr_pipeline"}, "('PIPELINES', 'document-vl-embedding', 'document-vl-embedding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/document_vl_embedding_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.multi_modal.document_vl_embedding_pipeline"}, "('PIPELINES', 'efficient-diffusion-tuning', 'efficient-diffusion-tuning')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/efficient_diffusion_tuning_pipeline.py", "imports": ["numpy", "torchvision", "typing", "cv2", "torch", "PIL"], "module": "modelscope.pipelines.multi_modal.efficient_diffusion_tuning_pipeline"}, "('PIPELINES', 'video-multi-modal-embedding', 'video-multi-modal-embedding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/video_multi_modal_embedding_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.multi_modal.video_multi_modal_embedding_pipeline"}, "('PIPELINES', 'video-temporal-grounding', 'soonet-video-temporal-grounding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/soonet_video_temporal_grounding_pipeline.py", "imports": ["numpy", "torchvision", "typing", "torch", "os"], "module": "modelscope.pipelines.multi_modal.soonet_video_temporal_grounding_pipeline"}, "('PIPELINES', 'image-captioning', 'image-captioning')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/image_captioning_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.multi_modal.image_captioning_pipeline"}, "('PIPELINES', 'video-question-answering', 'video-question-answering')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/video_question_answering_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.multi_modal.video_question_answering_pipeline"}, "('PIPELINES', 'ocr-recognition', 'ofa-ocr-recognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/ocr_recognition_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.multi_modal.ocr_recognition_pipeline"}, "('PIPELINES', 'image-text-retrieval', 'image-text-retrieval')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/image_text_retrieval_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.multi_modal.image_text_retrieval_pipeline"}, "('PIPELINES', 'text-to-image-synthesis', 'cones2-inference')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/cone2_pipeline/cones2_inference_pipeline.py", "imports": ["numpy", "math", "tqdm", "typing", "torch", "cv2", "PIL", "diffusers"], "module": "modelscope.pipelines.multi_modal.cone2_pipeline.cones2_inference_pipeline"}, "('PIPELINES', 'visual-entailment', 'visual-entailment')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/visual_entailment_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.multi_modal.visual_entailment_pipeline"}, "('PIPELINES', 'image-to-video', 'image-to-video-task-pipeline')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/image_to_video_pipeline.py", "imports": ["subprocess", "typing", "torch", "cv2", "tempfile", "einops", "os"], "module": "modelscope.pipelines.multi_modal.image_to_video_pipeline"}, "('PIPELINES', 'sudoku', 'ofa-sudoku')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/sudoku_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.multi_modal.sudoku_pipeline"}, "('PIPELINES', 'visual-question-answering', 'visual-question-answering')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/visual_question_answering_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.multi_modal.visual_question_answering_pipeline"}, "('PIPELINES', 'multi-modal-similarity', 'multi-modal-similarity')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/team_multi_modal_similarity_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.multi_modal.team_multi_modal_similarity_pipeline"}, "('PIPELINES', 'text2sql', 'ofa-text2sql')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/text2sql_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.multi_modal.text2sql_pipeline"}, "('PIPELINES', 'video-captioning', 'video-captioning')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/video_captioning_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.multi_modal.video_captioning_pipeline"}, "('PIPELINES', 'text-ranking', 'mgeo-ranking')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/mgeo_ranking_pipeline.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.pipelines.multi_modal.mgeo_ranking_pipeline"}, "('PIPELINES', 'image-text-retrieval', 'multi-modal-embedding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/multi_modal_embedding_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.multi_modal.multi_modal_embedding_pipeline"}, "('PIPELINES', 'multi-modal-embedding', 'multi-modal-embedding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/multi_modal_embedding_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.multi_modal.multi_modal_embedding_pipeline"}, "('PIPELINES', 'text-to-video-synthesis', 'videocomposer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/videocomposer_pipeline.py", "imports": ["numpy", "time", "torchvision", "functools", "subprocess", "mvextractor", "typing", "cv2", "torch", "PIL", "tempfile", "os", "random", "imageio"], "module": "modelscope.pipelines.multi_modal.videocomposer_pipeline"}, "('PIPELINES', 'multimodal-dialogue', 'multimodal-dialogue')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/multimodal_dialogue_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.multi_modal.multimodal_dialogue_pipeline"}, "('PIPELINES', 'text-to-image-synthesis', 'disco_guided_diffusion')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/disco_guided_diffusion_pipeline/disco_guided_diffusion.py", "imports": ["numpy", "gc", "torchvision", "math", "json", "cv2", "torch", "PIL", "importlib", "clip", "os"], "module": "modelscope.pipelines.multi_modal.disco_guided_diffusion_pipeline.disco_guided_diffusion"}, "('PIPELINES', 'generative-multi-modal-embedding', 'generative-multi-modal-embedding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/generative_multi_modal_embedding_pipeline.py", "imports": ["typing"], "module": "modelscope.pipelines.multi_modal.generative_multi_modal_embedding_pipeline"}, "('PIPELINES', 'text-to-image-synthesis', 'diffusers-stable-diffusion')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/diffusers_wrapped/stable_diffusion/stable_diffusion_pipeline.py", "imports": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "diffusers", "os"], "module": "modelscope.pipelines.multi_modal.diffusers_wrapped.stable_diffusion.stable_diffusion_pipeline"}, "('PIPELINES', 'text-to-image-synthesis', 'chinese-stable-diffusion')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/diffusers_wrapped/stable_diffusion/chinese_stable_diffusion_pipeline.py", "imports": ["numpy", "transformers", "typing", "torch", "cv2", "PIL", "diffusers"], "module": "modelscope.pipelines.multi_modal.diffusers_wrapped.stable_diffusion.chinese_stable_diffusion_pipeline"}, "('PIPELINES', 'video-to-video', 'video-to-video-pipeline')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/video_to_video_pipeline.py", "imports": ["subprocess", "typing", "torch", "cv2", "tempfile", "einops", "os"], "module": "modelscope.pipelines.multi_modal.video_to_video_pipeline"}, "('PIPELINES', 'text-to-video-synthesis', 'latent-text-to-video-synthesis')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/text_to_video_synthesis_pipeline.py", "imports": ["typing", "cv2", "torch", "tempfile", "einops", "os"], "module": "modelscope.pipelines.multi_modal.text_to_video_synthesis_pipeline"}, "('PIPELINES', 'visual-grounding', 'visual-grounding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/visual_grounding_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.multi_modal.visual_grounding_pipeline"}, "('PIPELINES', 'text-to-image-synthesis', 'text-to-image-synthesis')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/text_to_image_synthesis_pipeline.py", "imports": ["typing", "torch"], "module": "modelscope.pipelines.multi_modal.text_to_image_synthesis_pipeline"}, "('PIPELINES', 'visual-question-answering', 'gridvlp-multi-modal-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/gridvlp_pipeline.py", "imports": ["traceback", "numpy", "time", "transformers", "json", "typing", "torch", "PIL", "os"], "module": "modelscope.pipelines.multi_modal.gridvlp_pipeline"}, "('PIPELINES', 'multi-modal-embedding', 'gridvlp-multi-modal-embedding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/pipelines/multi_modal/gridvlp_pipeline.py", "imports": ["traceback", "numpy", "time", "transformers", "json", "typing", "torch", "PIL", "os"], "module": "modelscope.pipelines.multi_modal.gridvlp_pipeline"}, "('PREPROCESSORS', 'text-to-speech', 'kantts-data-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/tts.py", "imports": ["kantts", "os", "typing"], "module": "modelscope.preprocessors.tts"}, "('PREPROCESSORS', 'multi-modal', 'diffusion-image-generation-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/multi_modal.py", "imports": ["numpy", "torchvision", "re", "decord", "json", "timm", "typing", "torch", "PIL", "os", "io"], "module": "modelscope.preprocessors.multi_modal"}, "('PREPROCESSORS', 'multi-modal', 'ofa-tasks-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/multi_modal.py", "imports": ["numpy", "torchvision", "re", "decord", "json", "timm", "typing", "torch", "PIL", "os", "io"], "module": "modelscope.preprocessors.multi_modal"}, "('PREPROCESSORS', 'multi-modal', 'clip-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/multi_modal.py", "imports": ["numpy", "torchvision", "re", "decord", "json", "timm", "typing", "torch", "PIL", "os", "io"], "module": "modelscope.preprocessors.multi_modal"}, "('PREPROCESSORS', 'multi-modal', 'mplug-tasks-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/multi_modal.py", "imports": ["numpy", "torchvision", "re", "decord", "json", "timm", "typing", "torch", "PIL", "os", "io"], "module": "modelscope.preprocessors.multi_modal"}, "('PREPROCESSORS', 'multi-modal', 'vldoc-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/multi_modal.py", "imports": ["numpy", "torchvision", "re", "decord", "json", "timm", "typing", "torch", "PIL", "os", "io"], "module": "modelscope.preprocessors.multi_modal"}, "('PREPROCESSORS', 'multi-modal', 'hitea-tasks-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/multi_modal.py", "imports": ["numpy", "torchvision", "re", "decord", "json", "timm", "typing", "torch", "PIL", "os", "io"], "module": "modelscope.preprocessors.multi_modal"}, "('PREPROCESSORS', 'multi-modal', 'mplug-owl-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/multi_modal.py", "imports": ["numpy", "torchvision", "re", "decord", "json", "timm", "typing", "torch", "PIL", "os", "io"], "module": "modelscope.preprocessors.multi_modal"}, "('PREPROCESSORS', 'multi-modal', 'image-captioning-clip-interrogator-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/multi_modal.py", "imports": ["numpy", "torchvision", "re", "decord", "json", "timm", "typing", "torch", "PIL", "os", "io"], "module": "modelscope.preprocessors.multi_modal"}, "('PREPROCESSORS', 'audio', 'LinearAECAndFbank')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/audio.py", "imports": ["numpy", "typing", "torch", "scipy", "os", "io"], "module": "modelscope.preprocessors.audio"}, "('PREPROCESSORS', 'nlp', 'text-gen-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/text_generation_preprocessor.py", "imports": ["numpy", "os", "typing", "torch"], "module": "modelscope.preprocessors.nlp.text_generation_preprocessor"}, "('PREPROCESSORS', 'nlp', 'text-gen-jieba-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/text_generation_preprocessor.py", "imports": ["numpy", "os", "typing", "torch"], "module": "modelscope.preprocessors.nlp.text_generation_preprocessor"}, "('PREPROCESSORS', 'nlp', 'sentence-piece')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/text_generation_preprocessor.py", "imports": ["numpy", "os", "typing", "torch"], "module": "modelscope.preprocessors.nlp.text_generation_preprocessor"}, "('PREPROCESSORS', 'nlp', 'text2text-gen-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/text_generation_preprocessor.py", "imports": ["numpy", "os", "typing", "torch"], "module": "modelscope.preprocessors.nlp.text_generation_preprocessor"}, "('PREPROCESSORS', 'nlp', 'zero-shot-cls-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/zero_shot_classification_preprocessor.py", "imports": ["typing"], "module": "modelscope.preprocessors.nlp.zero_shot_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'thai-ner-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/token_classification_thai_preprocessor.py", "imports": ["typing"], "module": "modelscope.preprocessors.nlp.token_classification_thai_preprocessor"}, "('PREPROCESSORS', 'nlp', 'thai-wseg-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/token_classification_thai_preprocessor.py", "imports": ["typing"], "module": "modelscope.preprocessors.nlp.token_classification_thai_preprocessor"}, "('PREPROCESSORS', 'nlp', 're-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/relation_extraction_preprocessor.py", "imports": ["transformers", "typing"], "module": "modelscope.preprocessors.nlp.relation_extraction_preprocessor"}, "('PREPROCESSORS', 'nlp', 'fill-mask')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/fill_mask_preprocessor.py", "imports": ["numpy", "abc", "re", "typing", "torch", "os"], "module": "modelscope.preprocessors.nlp.fill_mask_preprocessor"}, "('PREPROCESSORS', 'nlp', 'fill-mask-ponet')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/fill_mask_preprocessor.py", "imports": ["numpy", "abc", "re", "typing", "torch", "os"], "module": "modelscope.preprocessors.nlp.fill_mask_preprocessor"}, "('PREPROCESSORS', 'nlp', 'conversational-text-to-sql')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/space_T_en/conversational_text_to_sql_preprocessor.py", "imports": ["json", "typing", "torch", "text2sql_lgesql", "os"], "module": "modelscope.preprocessors.nlp.space_T_en.conversational_text_to_sql_preprocessor"}, "('PREPROCESSORS', 'nlp', 'word-alignment')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/word_alignment_preprocessor.py", "imports": ["numpy", "typing", "torch", "itertools", "os"], "module": "modelscope.preprocessors.nlp.word_alignment_preprocessor"}, "('PREPROCESSORS', 'nlp', 'table-question-answering-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/space_T_cn/table_question_answering_preprocessor.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.preprocessors.nlp.space_T_cn.table_question_answering_preprocessor"}, "('PREPROCESSORS', 'nlp', 'dialog-state-tracking-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/space/dialog_state_tracking_preprocessor.py", "imports": ["typing"], "module": "modelscope.preprocessors.nlp.space.dialog_state_tracking_preprocessor"}, "('PREPROCESSORS', 'nlp', 'dialog-modeling-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/space/dialog_modeling_preprocessor.py", "imports": ["os", "typing"], "module": "modelscope.preprocessors.nlp.space.dialog_modeling_preprocessor"}, "('PREPROCESSORS', 'nlp', 'dialog-intent-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/space/dialog_intent_prediction_preprocessor.py", "imports": ["typing", "json", "os"], "module": "modelscope.preprocessors.nlp.space.dialog_intent_prediction_preprocessor"}, "('PREPROCESSORS', 'nlp', 'document-grounded-dialog-retrieval')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/document_grounded_dialog_retrieval_preprocessor.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.preprocessors.nlp.document_grounded_dialog_retrieval_preprocessor"}, "('PREPROCESSORS', 'nlp', 'document-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/document_segmentation_preprocessor.py", "imports": ["typing"], "module": "modelscope.preprocessors.nlp.document_segmentation_preprocessor"}, "('PREPROCESSORS', 'nlp', 'dialog-use-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/dialog_classification_use_preprocessor.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.preprocessors.nlp.dialog_classification_use_preprocessor"}, "('PREPROCESSORS', 'nlp', 'canmt-translation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/canmt_translation.py", "imports": ["typing", "sacremoses", "torch", "subword_nmt", "jieba", "os"], "module": "modelscope.preprocessors.nlp.canmt_translation"}, "('PREPROCESSORS', 'nlp', 'document-grounded-dialog-rerank')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/document_grounded_dialog_rerank_preprocessor.py", "imports": ["transformers", "typing", "torch", "copy", "os"], "module": "modelscope.preprocessors.nlp.document_grounded_dialog_rerank_preprocessor"}, "('PREPROCESSORS', 'nlp', 'nli-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/text_classification_preprocessor.py", "imports": ["numpy", "typing"], "module": "modelscope.preprocessors.nlp.text_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'sen-sim-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/text_classification_preprocessor.py", "imports": ["numpy", "typing"], "module": "modelscope.preprocessors.nlp.text_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'bert-seq-cls-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/text_classification_preprocessor.py", "imports": ["numpy", "typing"], "module": "modelscope.preprocessors.nlp.text_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'sen-cls-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/text_classification_preprocessor.py", "imports": ["numpy", "typing"], "module": "modelscope.preprocessors.nlp.text_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'Tokenize')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/bert_seq_cls_tokenizer.py", "imports": ["transformers", "typing"], "module": "modelscope.preprocessors.nlp.bert_seq_cls_tokenizer"}, "('PREPROCESSORS', 'nlp', 'word-segment-text-to-label-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/token_classification_preprocessor.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.preprocessors.nlp.token_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'ner-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/token_classification_preprocessor.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.preprocessors.nlp.token_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'token-cls-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/token_classification_preprocessor.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.preprocessors.nlp.token_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'sequence-labeling-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/token_classification_preprocessor.py", "imports": ["numpy", "typing", "torch"], "module": "modelscope.preprocessors.nlp.token_classification_preprocessor"}, "('PREPROCESSORS', 'nlp', 'machine-reading-comprehension-for-ner')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/machine_reading_comprehension_preprocessor.py", "imports": ["transformers", "os", "torch"], "module": "modelscope.preprocessors.nlp.machine_reading_comprehension_preprocessor"}, "('PREPROCESSORS', 'nlp', 'faq-question-answering-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/faq_question_answering_preprocessor.py", "imports": ["typing", "torch"], "module": "modelscope.preprocessors.nlp.faq_question_answering_preprocessor"}, "('PREPROCESSORS', 'nlp', 'text-ranking')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/text_ranking_preprocessor.py", "imports": ["transformers", "typing"], "module": "modelscope.preprocessors.nlp.text_ranking_preprocessor"}, "('PREPROCESSORS', 'nlp', 'viet-ner-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/token_classification_viet_preprocessor.py", "imports": ["typing", "torch"], "module": "modelscope.preprocessors.nlp.token_classification_viet_preprocessor"}, "('PREPROCESSORS', 'nlp', 'sentence-embedding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/sentence_embedding_preprocessor.py", "imports": ["typing"], "module": "modelscope.preprocessors.nlp.sentence_embedding_preprocessor"}, "('PREPROCESSORS', 'nlp', 'feature-extraction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/feature_extraction_preprocessor.py", "imports": ["numpy", "typing"], "module": "modelscope.preprocessors.nlp.feature_extraction_preprocessor"}, "('PREPROCESSORS', 'nlp', 'mgeo-ranking')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/mgeo_ranking_preprocessor.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.preprocessors.nlp.mgeo_ranking_preprocessor"}, "('PREPROCESSORS', 'nlp', 'text-error-correction')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/text_error_correction.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.preprocessors.nlp.text_error_correction"}, "('PREPROCESSORS', 'nlp', 'translation-evaluation-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/translation_evaluation_preprocessor.py", "imports": ["transformers", "typing", "torch"], "module": "modelscope.preprocessors.nlp.translation_evaluation_preprocessor"}, "('PREPROCESSORS', 'nlp', 'mglm-summarization')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/mglm_summarization_preprocessor.py", "imports": ["re", "os", "typing"], "module": "modelscope.preprocessors.nlp.mglm_summarization_preprocessor"}, "('PREPROCESSORS', 'nlp', 'siamese-uie-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/siamese_uie_preprocessor.py", "imports": ["transformers", "typing"], "module": "modelscope.preprocessors.nlp.siamese_uie_preprocessor"}, "('PREPROCESSORS', 'nlp', 'document-grounded-dialog-generate')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/nlp/document_grounded_dialog_generate_preprocessor.py", "imports": ["transformers", "os", "typing", "torch"], "module": "modelscope.preprocessors.nlp.document_grounded_dialog_generate_preprocessor"}, "('PREPROCESSORS', 'cv', 'load-image')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/image.py", "imports": ["numpy", "typing", "cv2", "PIL", "io"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'object-detection-tinynas-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/image.py", "imports": ["numpy", "typing", "cv2", "PIL", "io"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'image-color-enhance-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/image.py", "imports": ["numpy", "typing", "cv2", "PIL", "io"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'image-denoise-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/image.py", "imports": ["numpy", "typing", "cv2", "PIL", "io"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'image-deblur-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/image.py", "imports": ["numpy", "typing", "cv2", "PIL", "io"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'image-portrait-enhancement-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/image.py", "imports": ["numpy", "typing", "cv2", "PIL", "io"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'image-instance-segmentation-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/image.py", "imports": ["numpy", "typing", "cv2", "PIL", "io"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'video-summarization-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/image.py", "imports": ["numpy", "typing", "cv2", "PIL", "io"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'cv', 'image-classification-bypass-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/image.py", "imports": ["numpy", "typing", "cv2", "PIL", "io"], "module": "modelscope.preprocessors.image"}, "('PREPROCESSORS', 'audio', 'sen-cls-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/speaker.py", "imports": ["typing", "torch"], "module": "modelscope.preprocessors.speaker"}, "('PREPROCESSORS', 'audio', 'token-cls-tokenizer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/speaker.py", "imports": ["typing", "torch"], "module": "modelscope.preprocessors.speaker"}, "('PREPROCESSORS', 'audio', 'wav-to-lists')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/kws.py", "imports": ["yaml", "os", "typing"], "module": "modelscope.preprocessors.kws"}, "('PREPROCESSORS', 'audio', 'wav-to-scp')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/asr.py", "imports": ["os", "typing"], "module": "modelscope.preprocessors.asr"}, "('PREPROCESSORS', 'science', 'unifold-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/science/uni_fold.py", "imports": ["numpy", "gzip", "ipdb", "tarfile", "hashlib", "requests", "random", "time", "unittest", "re", "pickle", "json", "tqdm", "typing", "torch", "pathlib", "logging", "os"], "module": "modelscope.preprocessors.science.uni_fold"}, "('PREPROCESSORS', 'default', 'Compose')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/common.py", "imports": ["numpy", "collections", "time", "typing", "torch"], "module": "modelscope.preprocessors.common"}, "('PREPROCESSORS', 'default', 'ToTensor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/common.py", "imports": ["numpy", "collections", "time", "typing", "torch"], "module": "modelscope.preprocessors.common"}, "('PREPROCESSORS', 'default', 'Filter')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/common.py", "imports": ["numpy", "collections", "time", "typing", "torch"], "module": "modelscope.preprocessors.common"}, "('PREPROCESSORS', 'default', 'ToNumpy')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/common.py", "imports": ["numpy", "collections", "time", "typing", "torch"], "module": "modelscope.preprocessors.common"}, "('PREPROCESSORS', 'default', 'Rename')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/common.py", "imports": ["numpy", "collections", "time", "typing", "torch"], "module": "modelscope.preprocessors.common"}, "('PREPROCESSORS', 'default', 'Identity')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/common.py", "imports": ["numpy", "collections", "time", "typing", "torch"], "module": "modelscope.preprocessors.common"}, "('PREPROCESSORS', 'cv', 'controllable-image-generation-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/cv/controllable_image_generation.py", "imports": ["numpy", "torchvision", "math", "typing", "torch", "cv2", "PIL", "os"], "module": "modelscope.preprocessors.cv.controllable_image_generation"}, "('PREPROCESSORS', 'cv', 'bad-image-detecting-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/cv/bad_image_detecting_preprocessor.py", "imports": ["numpy", "torchvision", "math", "typing", "torch", "PIL"], "module": "modelscope.preprocessors.cv.bad_image_detecting_preprocessor"}, "('PREPROCESSORS', 'cv', 'RandomCrop')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "os"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'RandomResizedCrop')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "os"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'Resize')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "os"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'CenterCrop')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "os"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'RandomHorizontalFlip')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "os"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'Normalize')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "os"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'ImageToTensor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "os"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'image-classification-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/cv/image_classification_preprocessor.py", "imports": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "os"], "module": "modelscope.preprocessors.cv.image_classification_preprocessor"}, "('PREPROCESSORS', 'cv', 'image-quality_assessment-man-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/cv/image_quality_assessment_man.py", "imports": ["numpy", "torchvision", "math", "typing", "torch", "PIL"], "module": "modelscope.preprocessors.cv.image_quality_assessment_man"}, "('PREPROCESSORS', 'cv', 'image-quality_assessment-mos-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/cv/image_quality_assessment_mos.py", "imports": ["numpy", "torchvision", "math", "typing", "cv2"], "module": "modelscope.preprocessors.cv.image_quality_assessment_mos"}, "('PREPROCESSORS', 'cv', 'image-classification-mmcv-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/cv/mmcls_preprocessor.py", "imports": ["numpy", "os", "typing"], "module": "modelscope.preprocessors.cv.mmcls_preprocessor"}, "('PREPROCESSORS', 'cv', 'image-demoire-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/cv/image_restoration_preprocessor.py", "imports": ["numpy", "torchvision", "math", "typing", "torch", "PIL"], "module": "modelscope.preprocessors.cv.image_restoration_preprocessor"}, "('PREPROCESSORS', 'cv', 'movie-scene-segmentation-preprocessor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/preprocessors/video.py", "imports": ["numpy", "torchvision", "decord", "math", "torch", "urllib", "uuid", "tempfile", "os", "random"], "module": "modelscope.preprocessors.video"}, "('PARALLEL', 'default', 'DistributedDataParallel')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/parallel/builder.py", "imports": ["torch"], "module": "modelscope.trainers.parallel.builder"}, "('LR_SCHEDULER', 'default', 'ConstantWarmup')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/lrscheduler/warmup/warmup.py", "imports": [], "module": "modelscope.trainers.lrscheduler.warmup.warmup"}, "('LR_SCHEDULER', 'default', 'LinearWarmup')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/lrscheduler/warmup/warmup.py", "imports": [], "module": "modelscope.trainers.lrscheduler.warmup.warmup"}, "('LR_SCHEDULER', 'default', 'ExponentialWarmup')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/lrscheduler/warmup/warmup.py", "imports": [], "module": "modelscope.trainers.lrscheduler.warmup.warmup"}, "('TRAINERS', 'default', 'dummy')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/base.py", "imports": ["abc", "time", "os", "typing"], "module": "modelscope.trainers.base"}, "('HOOKS', 'default', 'MegatronHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/distributed/megatron_hook.py", "imports": ["megatron_util", "os", "shutil", "torch"], "module": "modelscope.trainers.hooks.distributed.megatron_hook"}, "('HOOKS', 'default', 'DeepspeedHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/distributed/deepspeed_hook.py", "imports": ["transformers", "deepspeed", "functools", "math", "megatron_util", "shutil", "torch", "os"], "module": "modelscope.trainers.hooks.distributed.deepspeed_hook"}, "('HOOKS', 'default', 'DDPHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/distributed/ddp_hook.py", "imports": [], "module": "modelscope.trainers.hooks.distributed.ddp_hook"}, "('HOOKS', 'default', 'IterTimerHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/iter_timer_hook.py", "imports": ["time"], "module": "modelscope.trainers.hooks.iter_timer_hook"}, "('HOOKS', 'default', 'EarlyStopHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/early_stop_hook.py", "imports": ["numpy"], "module": "modelscope.trainers.hooks.early_stop_hook"}, "('HOOKS', 'default', 'EvaluationHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/evaluation_hook.py", "imports": ["collections", "typing"], "module": "modelscope.trainers.hooks.evaluation_hook"}, "('HOOKS', 'default', 'LoadCheckpointHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/checkpoint/load_checkpoint_hook.py", "imports": ["numpy", "packaging", "typing", "torch", "random"], "module": "modelscope.trainers.hooks.checkpoint.load_checkpoint_hook"}, "('HOOKS', 'default', 'CheckpointHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/checkpoint/checkpoint_hook.py", "imports": ["numpy", "json", "shutil", "torch", "typing", "os", "random"], "module": "modelscope.trainers.hooks.checkpoint.checkpoint_hook"}, "('HOOKS', 'default', 'BestCkptSaverHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/checkpoint/checkpoint_hook.py", "imports": ["numpy", "json", "shutil", "torch", "typing", "os", "random"], "module": "modelscope.trainers.hooks.checkpoint.checkpoint_hook"}, "('HOOKS', 'default', 'SparsityHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/compression/sparsity_hook.py", "imports": ["os"], "module": "modelscope.trainers.hooks.compression.sparsity_hook"}, "('HOOKS', 'default', 'ApexAMPOptimizerHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/optimizer/apex_optimizer_hook.py", "imports": ["packaging", "logging", "torch"], "module": "modelscope.trainers.hooks.optimizer.apex_optimizer_hook"}, "('HOOKS', 'default', 'OptimizerHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/optimizer/base.py", "imports": ["logging", "torch"], "module": "modelscope.trainers.hooks.optimizer.base"}, "('HOOKS', 'default', 'NoneOptimizerHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/optimizer/base.py", "imports": ["logging", "torch"], "module": "modelscope.trainers.hooks.optimizer.base"}, "('HOOKS', 'default', 'TorchAMPOptimizerHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/optimizer/torch_optimizer_hook.py", "imports": ["logging"], "module": "modelscope.trainers.hooks.optimizer.torch_optimizer_hook"}, "('HOOKS', 'default', 'LrSchedulerHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/lr_scheduler_hook.py", "imports": [], "module": "modelscope.trainers.hooks.lr_scheduler_hook"}, "('HOOKS', 'default', 'PlateauLrSchedulerHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/lr_scheduler_hook.py", "imports": [], "module": "modelscope.trainers.hooks.lr_scheduler_hook"}, "('HOOKS', 'default', 'NoneLrSchedulerHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/lr_scheduler_hook.py", "imports": [], "module": "modelscope.trainers.hooks.lr_scheduler_hook"}, "('HOOKS', 'default', 'ClipClampLogitScaleHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/clip_clamp_logit_scale_hook.py", "imports": ["torch"], "module": "modelscope.trainers.hooks.clip_clamp_logit_scale_hook"}, "('HOOKS', 'default', 'TextLoggerHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/logger/text_logger_hook.py", "imports": ["collections", "json", "torch", "datetime", "os"], "module": "modelscope.trainers.hooks.logger.text_logger_hook"}, "('HOOKS', 'default', 'TensorboardHook')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/hooks/logger/tensorboard_hook.py", "imports": ["numpy", "os", "torch"], "module": "modelscope.trainers.hooks.logger.tensorboard_hook"}, "('TRAINERS', 'default', 'nlp-base-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp_trainer.py", "imports": ["numpy", "os", "typing", "torch"], "module": "modelscope.trainers.nlp_trainer"}, "('TRAINERS', 'default', 'nlp-veco-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp_trainer.py", "imports": ["numpy", "os", "typing", "torch"], "module": "modelscope.trainers.nlp_trainer"}, "('TRAINERS', 'default', 'faq-question-answering-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/faq_question_answering_trainer.py", "imports": ["numpy", "collections", "dataclasses", "functools", "contextlib", "torch", "typing", "distutils"], "module": "modelscope.trainers.nlp.faq_question_answering_trainer"}, "('TRAINERS', 'default', 'siamese-uie-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/siamese_uie_trainer.py", "imports": ["numpy", "collections", "time", "math", "json", "typing", "torch", "random", "os"], "module": "modelscope.trainers.nlp.siamese_uie_trainer"}, "('TRAINERS', 'default', 'nlp-plug-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/plug_trainer.py", "imports": ["deepspeed", "megatron_util", "typing", "torch", "os"], "module": "modelscope.trainers.nlp.plug_trainer"}, "('TRAINERS', 'default', 'document-grounded-dialog-generate-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/document_grounded_dialog_generate_trainer.py", "imports": ["collections", "transformers", "re", "tqdm", "json", "sacrebleu", "string", "torch", "rouge", "os"], "module": "modelscope.trainers.nlp.document_grounded_dialog_generate_trainer"}, "('TRAINERS', 'default', 'nlp-gpt-moe-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/gpt_moe_trainer.py", "imports": ["collections", "megatron_util", "typing", "torch", "os"], "module": "modelscope.trainers.nlp.gpt_moe_trainer"}, "('TRAINERS', 'default', 'document-grounded-dialog-retrieval-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/document_grounded_dialog_retrieval_trainer.py", "imports": ["numpy", "transformers", "tqdm", "json", "torch", "faiss", "os"], "module": "modelscope.trainers.nlp.document_grounded_dialog_retrieval_trainer"}, "('TRAINERS', 'default', 'dialog-modeling-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/space/dialog_modeling_trainer.py", "imports": ["numpy", "time", "os", "typing"], "module": "modelscope.trainers.nlp.space.dialog_modeling_trainer"}, "('TRAINERS', 'default', 'dialog-intent-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/space/dialog_intent_trainer.py", "imports": ["numpy", "os", "typing"], "module": "modelscope.trainers.nlp.space.dialog_intent_trainer"}, "('TRAINERS', 'default', 'nlp-text-ranking-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/text_ranking_trainer.py", "imports": ["numpy", "time", "dataclasses", "tqdm", "typing", "torch"], "module": "modelscope.trainers.nlp.text_ranking_trainer"}, "('TRAINERS', 'default', 'translation-evaluation-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/translation_evaluation_trainer.py", "imports": ["transformers", "math", "tqdm", "pandas", "typing", "torch", "os", "random"], "module": "modelscope.trainers.nlp.translation_evaluation_trainer"}, "('TRAINERS', 'default', 'table-question-answering-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/table_question_answering_trainer.py", "imports": ["numpy", "time", "tqdm", "json", "typing", "torch", "os"], "module": "modelscope.trainers.nlp.table_question_answering_trainer"}, "('TRAINERS', 'default', 'document-grounded-dialog-rerank-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/document_grounded_dialog_rerank_trainer.py", "imports": ["numpy", "time", "transformers", "typing", "torch", "os", "random"], "module": "modelscope.trainers.nlp.document_grounded_dialog_rerank_trainer"}, "('TRAINERS', 'default', 'nlp-gpt3-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/gpt3_trainer.py", "imports": ["copy", "os", "typing", "torch"], "module": "modelscope.trainers.nlp.gpt3_trainer"}, "('TRAINERS', 'default', 'text-generation-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/text_generation_trainer.py", "imports": ["typing", "torch"], "module": "modelscope.trainers.nlp.text_generation_trainer"}, "('TRAINERS', 'default', 'bert-sentiment-analysis')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/sequence_classification_trainer.py", "imports": ["numpy", "time", "typing"], "module": "modelscope.trainers.nlp.sequence_classification_trainer"}, "('TRAINERS', 'default', 'csanmt-translation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/csanmt_translation_trainer.py", "imports": ["tensorflow", "time", "os", "typing"], "module": "modelscope.trainers.nlp.csanmt_translation_trainer"}, "('TRAINERS', 'default', 'nlp-sentence-embedding-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/nlp/sentence_embedding_trainer.py", "imports": ["numpy", "time", "transformers", "dataclasses", "tqdm", "typing", "torch"], "module": "modelscope.trainers.nlp.sentence_embedding_trainer"}, "('TRAINERS', 'default', 'trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/trainer.py", "imports": ["collections", "functools", "json", "inspect", "torch", "typing", "distutils", "copy", "os"], "module": "modelscope.trainers.trainer"}, "('TRAINERS', 'default', 'speech-asr-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/audio/asr_trainer.py", "imports": ["json", "shutil", "typing", "tempfile", "funasr", "os"], "module": "modelscope.trainers.audio.asr_trainer"}, "('TRAINERS', 'default', 'speech_kws_fsmn_char_ctc_nearfield')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/audio/kws_nearfield_trainer.py", "imports": ["yaml", "re", "typing", "torch", "tensorboardX", "datetime", "copy", "os"], "module": "modelscope.trainers.audio.kws_nearfield_trainer"}, "('TRAINERS', 'default', 'speech-kantts-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/audio/tts_trainer.py", "imports": ["json", "typing", "shutil", "tempfile", "zipfile", "os"], "module": "modelscope.trainers.audio.tts_trainer"}, "('TRAINERS', 'default', 'speech_dfsmn_kws_char_farfield')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/audio/kws_farfield_trainer.py", "imports": ["numpy", "math", "pickle", "typing", "torch", "datetime", "os", "glob"], "module": "modelscope.trainers.audio.kws_farfield_trainer"}, "('TRAINERS', 'default', 'speech-separation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/audio/separation_trainer.py", "imports": ["numpy", "tqdm", "typing", "torch", "speechbrain", "csv", "os", "torchaudio"], "module": "modelscope.trainers.audio.separation_trainer"}, "('TRAINERS', 'default', 'speech_frcrn_ans_cirm_16k')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/audio/ans_trainer.py", "imports": [], "module": "modelscope.trainers.audio.ans_trainer"}, "('TRAINERS', 'default', 'ocr-recognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/ocr_recognition_trainer.py", "imports": ["collections", "time", "torch"], "module": "modelscope.trainers.cv.ocr_recognition_trainer"}, "('TRAINERS', 'default', 'image-portrait-enhancement')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/image_portrait_enhancement_trainer.py", "imports": ["collections", "torch"], "module": "modelscope.trainers.cv.image_portrait_enhancement_trainer"}, "('TRAINERS', 'default', 'ocr-detection-db')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/ocr_detection_db_trainer.py", "imports": ["numpy", "time", "math", "tqdm", "typing", "torch", "easydict", "datetime", "copy", "os"], "module": "modelscope.trainers.cv.ocr_detection_db_trainer"}, "('TRAINERS', 'default', 'action-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/action_detection_trainer.py", "imports": ["typing", "torch", "fvcore", "detectron2", "os"], "module": "modelscope.trainers.cv.action_detection_trainer"}, "('TRAINERS', 'default', 'tinynas-damoyolo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/image_detection_damoyolo_trainer.py", "imports": ["time", "math", "typing", "torch", "easydict", "datetime", "os"], "module": "modelscope.trainers.cv.image_detection_damoyolo_trainer"}, "('TRAINERS', 'default', 'face-detection-scrfd')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/face_detection_scrfd_trainer.py", "imports": ["time", "os", "typing", "copy"], "module": "modelscope.trainers.cv.face_detection_scrfd_trainer"}, "('TRAINERS', 'default', 'image-instance-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/image_instance_segmentation_trainer.py", "imports": [], "module": "modelscope.trainers.cv.image_instance_segmentation_trainer"}, "('TRAINERS', 'default', 'image-inpainting')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/image_inpainting_trainer.py", "imports": ["collections", "time", "torch"], "module": "modelscope.trainers.cv.image_inpainting_trainer"}, "('TRAINERS', 'default', 'nerf-recon-acc')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/nerf_recon_acc_trainer.py", "imports": ["numpy", "time", "tqdm", "typing", "cv2", "torch", "datetime", "os", "random", "glob"], "module": "modelscope.trainers.cv.nerf_recon_acc_trainer"}, "('TRAINERS', 'default', 'cartoon-translation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/cartoon_translation_trainer.py", "imports": ["numpy", "packaging", "tqdm", "typing", "os", "tensorflow"], "module": "modelscope.trainers.cv.cartoon_translation_trainer"}, "('TRAINERS', 'default', 'image-fewshot-detection')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/image_defrcn_fewshot_detection_trainer.py", "imports": ["collections", "typing", "torch", "detectron2", "os"], "module": "modelscope.trainers.cv.image_defrcn_fewshot_detection_trainer"}, "('TRAINERS', 'default', 'image-classification')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/image_classifition_trainer.py", "imports": ["numpy", "time", "typing", "torch", "copy", "os"], "module": "modelscope.trainers.cv.image_classifition_trainer"}, "('TRAINERS', 'default', 'movie-scene-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/movie_scene_segmentation_trainer.py", "imports": [], "module": "modelscope.trainers.cv.movie_scene_segmentation_trainer"}, "('TRAINERS', 'default', 'vision-efficient-tuning')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/vision_efficient_tuning_trainer.py", "imports": ["typing", "torch"], "module": "modelscope.trainers.cv.vision_efficient_tuning_trainer"}, "('TRAINERS', 'default', 'card-detection-scrfd')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/card_detection_scrfd_trainer.py", "imports": [], "module": "modelscope.trainers.cv.card_detection_scrfd_trainer"}, "('TRAINERS', 'default', 'referring-video-object-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/cv/referring_video_object_segmentation_trainer.py", "imports": ["os", "torch"], "module": "modelscope.trainers.cv.referring_video_object_segmentation_trainer"}, "('TRAINERS', 'default', 'dreambooth-diffusion')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/multi_modal/dreambooth_diffusion/dreambooth_diffusion_trainer.py", "imports": ["collections", "torchvision", "tqdm", "warnings", "shutil", "torch", "typing", "PIL", "itertools", "pathlib", "hashlib", "diffusers"], "module": "modelscope.trainers.multi_modal.dreambooth_diffusion.dreambooth_diffusion_trainer"}, "('TRAINERS', 'default', 'lora-diffusion')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/multi_modal/lora_diffusion/lora_diffusion_trainer.py", "imports": ["diffusers", "typing", "torch"], "module": "modelscope.trainers.multi_modal.lora_diffusion.lora_diffusion_trainer"}, "('TRAINERS', 'default', 'image-classification-team')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/multi_modal/team/team_trainer.py", "imports": ["numpy", "collections", "typing", "torch", "sklearn", "os"], "module": "modelscope.trainers.multi_modal.team.team_trainer"}, "('TRAINERS', 'default', 'cones2-inference')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/multi_modal/cones2/cones_trainer.py", "imports": ["torchvision", "torch", "PIL", "pathlib", "diffusers"], "module": "modelscope.trainers.multi_modal.cones2.cones_trainer"}, "('TRAINERS', 'default', 'custom-diffusion')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/multi_modal/custom_diffusion/custom_diffusion_trainer.py", "imports": ["numpy", "torchvision", "tqdm", "json", "warnings", "torch", "typing", "PIL", "itertools", "pathlib", "os", "hashlib", "diffusers", "random"], "module": "modelscope.trainers.multi_modal.custom_diffusion.custom_diffusion_trainer"}, "('TRAINERS', 'default', 'stable-diffusion')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/multi_modal/stable_diffusion/stable_diffusion_trainer.py", "imports": ["typing", "torch"], "module": "modelscope.trainers.multi_modal.stable_diffusion.stable_diffusion_trainer"}, "('TRAINERS', 'default', 'clip-multi-modal-embedding')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/multi_modal/clip/clip_trainer.py", "imports": ["math", "os", "typing", "torch"], "module": "modelscope.trainers.multi_modal.clip.clip_trainer"}, "('TRAINERS', 'default', 'lora-diffusion-xl')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/multi_modal/lora_diffusion_xl/lora_diffusion_xl_trainer.py", "imports": ["diffusers", "typing", "torch"], "module": "modelscope.trainers.multi_modal.lora_diffusion_xl.lora_diffusion_xl_trainer"}, "('TRAINERS', 'default', 'efficient-diffusion-tuning')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/multi_modal/efficient_diffusion_tuning/efficient_diffusion_tuning_trainer.py", "imports": ["typing", "torch"], "module": "modelscope.trainers.multi_modal.efficient_diffusion_tuning.efficient_diffusion_tuning_trainer"}, "('TRAINERS', 'default', 'ofa')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/multi_modal/ofa/ofa_trainer.py", "imports": ["functools", "math", "json", "shutil", "torch", "typing", "tempfile", "os"], "module": "modelscope.trainers.multi_modal.ofa.ofa_trainer"}, "('TRAINERS', 'default', 'mplug')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/multi_modal/mplug/mplug_trainer.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.trainers.multi_modal.mplug.mplug_trainer"}, "('TRAINERS', 'default', 'mgeo-ranking-trainer')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/trainers/multi_modal/mgeo_ranking_trainer.py", "imports": ["dataclasses", "typing", "torch"], "module": "modelscope.trainers.multi_modal.mgeo_ranking_trainer"}, "('CUSTOM_DATASETS', 'image-colorization', 'ddcolor')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_colorization/image_colorization_dataset.py", "imports": ["numpy", "cv2", "torch"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.image_colorization.image_colorization_dataset"}, "('CUSTOM_DATASETS', 'language-guided-video-summarization', 'clip-it-language-guided-video-summarization')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/language_guided_video_summarization_dataset.py", "imports": ["numpy", "h5py", "json", "torch", "os"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.language_guided_video_summarization_dataset"}, "('CUSTOM_DATASETS', 'image-quality-assessment-degradation', 'image-quality-assessment-degradation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_quality_assessment_degradation/image_quality_assessment_degradation_dataset.py", "imports": ["torchvision"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.image_quality_assessment_degradation.image_quality_assessment_degradation_dataset"}, "('CUSTOM_DATASETS', 'bad-image-detecting', 'bad-image-detecting')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/bad_image_detecting/bad_image_detecting_dataset.py", "imports": [], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.bad_image_detecting.bad_image_detecting_dataset"}, "('CUSTOM_DATASETS', 'video-super-resolution', 'real-basicvsr')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/video_super_resolution/video_super_resolution_dataset.py", "imports": ["numpy", "collections", "cv2", "torch"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.video_super_resolution.video_super_resolution_dataset"}, "('CUSTOM_DATASETS', 'image-segmentation', 'cascade_mask_rcnn_swin')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_instance_segmentation_coco_dataset.py", "imports": ["numpy", "os", "pycocotools"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.image_instance_segmentation_coco_dataset"}, "('CUSTOM_DATASETS', 'ocr-recognition', 'OCRRecognition')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/ocr_recognition_dataset.py", "imports": ["numpy", "json", "cv2", "torch", "PIL", "lmdb", "os", "six"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_recognition_dataset"}, "('CUSTOM_DATASETS', 'video-frame-interpolation', 'video-frame-interpolation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/video_frame_interpolation/video_frame_interpolation_dataset.py", "imports": ["numpy", "cv2", "torch"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.video_frame_interpolation.video_frame_interpolation_dataset"}, "('CUSTOM_DATASETS', 'image-deblurring', 'RedsDataset')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/reds_image_deblurring_dataset.py", "imports": ["numpy", "cv2"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.reds_image_deblurring_dataset"}, "('CUSTOM_DATASETS', 'image-portrait-enhancement', 'PairedDataset')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_portrait_enhancement/image_portrait_enhancement_dataset.py", "imports": ["numpy", "cv2"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.image_portrait_enhancement.image_portrait_enhancement_dataset"}, "('CUSTOM_DATASETS', 'image-inpainting', 'FFTInpainting')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_inpainting/image_inpainting_dataset.py", "imports": ["numpy", "enum", "cv2", "albumentations", "os", "glob"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.image_inpainting.image_inpainting_dataset"}, "('CUSTOM_DATASETS', 'text-ranking', 'mgeo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/mgeo_ranking_dataset.py", "imports": ["random", "json", "typing", "torch"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.mgeo_ranking_dataset"}, "('CUSTOM_DATASETS', 'movie-scene-segmentation', 'resnet50-bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/movie_scene_segmentation/movie_scene_segmentation_dataset.py", "imports": ["torchvision", "json", "torch", "copy", "random", "os"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.movie_scene_segmentation.movie_scene_segmentation_dataset"}, "('CUSTOM_DATASETS', 'image-deblurring', 'GoproDataset')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/gopro_image_deblurring_dataset.py", "imports": ["numpy", "cv2"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.gopro_image_deblurring_dataset"}, "('CUSTOM_DATASETS', 'text-ranking', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/text_ranking_dataset.py", "imports": ["random", "typing", "torch"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.text_ranking_dataset"}, "('CUSTOM_DATASETS', 'sentence-embedding', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/text_ranking_dataset.py", "imports": ["random", "typing", "torch"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.text_ranking_dataset"}, "('CUSTOM_DATASETS', 'image-denoising', 'SiddDataset')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/sidd_image_denoising/sidd_image_denoising_dataset.py", "imports": ["numpy", "cv2"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.sidd_image_denoising.sidd_image_denoising_dataset"}, "('CUSTOM_DATASETS', 'image-quality-assessment-mos', 'image-quality-assessment-mos')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/image_quality_assmessment_mos/image_quality_assessment_mos_dataset.py", "imports": [], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.image_quality_assmessment_mos.image_quality_assessment_mos_dataset"}, "('CUSTOM_DATASETS', 'nli', 'veco')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/veco_dataset.py", "imports": ["numpy", "datasets", "typing"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.veco_dataset"}, "('CUSTOM_DATASETS', 'referring-video-object-segmentation', 'swinT-referring-video-object-segmentation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/referring_video_object_segmentation/referring_video_object_segmentation_dataset.py", "imports": ["numpy", "h5py", "torchvision", "tqdm", "json", "pandas", "torch", "pycocotools", "os", "glob"], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.referring_video_object_segmentation.referring_video_object_segmentation_dataset"}, "('CUSTOM_DATASETS', 'video-stabilization', 'video-stabilization')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/msdatasets/dataset_cls/custom_datasets/video_stabilization/video_stabilization_dataset.py", "imports": [], "module": "modelscope.msdatasets.dataset_cls.custom_datasets.video_stabilization.video_stabilization_dataset"}, "('EXPORTERS', 'translation', 'csanmt-translation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/csanmt_for_translation_exporter.py", "imports": ["tensorflow", "os", "typing"], "module": "modelscope.exporters.nlp.csanmt_for_translation_exporter"}, "('EXPORTERS', 'zero-shot-classification', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/sbert_for_zero_shot_classification_exporter.py", "imports": ["collections", "typing"], "module": "modelscope.exporters.nlp.sbert_for_zero_shot_classification_exporter"}, "('EXPORTERS', 'zero-shot-classification', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/sbert_for_zero_shot_classification_exporter.py", "imports": ["collections", "typing"], "module": "modelscope.exporters.nlp.sbert_for_zero_shot_classification_exporter"}, "('EXPORTERS', 'transformer-crf', 'transformer-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/model_for_token_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.model_for_token_classification_exporter"}, "('EXPORTERS', 'token-classification', 'transformer-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/model_for_token_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.model_for_token_classification_exporter"}, "('EXPORTERS', 'named-entity-recognition', 'transformer-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/model_for_token_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.model_for_token_classification_exporter"}, "('EXPORTERS', 'part-of-speech', 'transformer-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/model_for_token_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.model_for_token_classification_exporter"}, "('EXPORTERS', 'word-segmentation', 'transformer-crf')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/model_for_token_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.model_for_token_classification_exporter"}, "('EXPORTERS', 'text-classification', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'text-classification', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'sentence-similarity', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'sentiment-classification', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'nli', 'bert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'sentence-similarity', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'sentiment-classification', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'nli', 'structbert')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/nlp/sbert_for_sequence_classification_exporter.py", "imports": ["collections", "typing", "torch"], "module": "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter"}, "('EXPORTERS', 'acoustic-noise-suppression', 'speech_dfsmn_ans')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/audio/ans_dfsmn_exporter.py", "imports": ["os", "torch"], "module": "modelscope.exporters.audio.ans_dfsmn_exporter"}, "('EXPORTERS', 'domain-specific-object-detection', 'tinynas-damoyolo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/cv/object_detection_damoyolo_exporter.py", "imports": ["numpy", "functools", "typing", "torch", "onnx", "os"], "module": "modelscope.exporters.cv.object_detection_damoyolo_exporter"}, "('EXPORTERS', 'image-object-detection', 'tinynas-damoyolo')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/cv/object_detection_damoyolo_exporter.py", "imports": ["numpy", "functools", "typing", "torch", "onnx", "os"], "module": "modelscope.exporters.cv.object_detection_damoyolo_exporter"}, "('EXPORTERS', 'default', 'cartoon-translation')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/cv/cartoon_translation_exporter.py", "imports": ["tensorflow", "packaging", "os", "typing"], "module": "modelscope.exporters.cv.cartoon_translation_exporter"}, "('EXPORTERS', 'face-detection', 'scrfd')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/cv/face_detection_scrfd_exporter.py", "imports": ["numpy", "functools", "typing", "torch", "onnx", "os"], "module": "modelscope.exporters.cv.face_detection_scrfd_exporter"}, "('EXPORTERS', 'text-to-image-synthesis', 'stable-diffusion')": {"filepath": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope/exporters/multi_modal/stable_diffusion_exporter.py", "imports": ["collections", "packaging", "argparse", "shutil", "torch", "typing", "pathlib", "onnx", "diffusers", "os"], "module": "modelscope.exporters.multi_modal.stable_diffusion_exporter"}}, "requirements": {"modelscope.models.base.base_head": ["abc", "typing"], "modelscope.models.base.base_model": ["abc", "os", "typing"], "modelscope.models.base.base_torch_model": ["functools", "packaging", "typing", "torch", "copy", "os"], "modelscope.models.base.base_torch_head": ["typing", "torch"], "modelscope.models.nlp.ponet.fill_mask": ["transformers", "torch"], "modelscope.models.nlp.ponet.document_segmentation": ["typing", "torch"], "modelscope.models.nlp.ponet.tokenization": ["transformers", "typing"], "modelscope.models.nlp.ponet.backbone": ["transformers", "math", "packaging", "torch", "distutils"], "modelscope.models.nlp.ponet.configuration": ["transformers"], "modelscope.models.nlp.heads.text_generation_head": ["typing", "torch"], "modelscope.models.nlp.heads.token_classification_head": ["typing", "torch"], "modelscope.models.nlp.heads.infromation_extraction_head": ["torch"], "modelscope.models.nlp.heads.text_classification_head": ["typing", "torch"], "modelscope.models.nlp.heads.torch_pretrain_head": ["transformers", "typing", "torch"], "modelscope.models.nlp.heads.text_ranking_head": ["typing", "torch"], "modelscope.models.nlp.heads.fill_mask_head": ["transformers", "typing", "torch"], "modelscope.models.nlp.heads.crf_head": ["transformers", "typing", "torch"], "modelscope.models.nlp.veco.fill_mask": ["transformers"], "modelscope.models.nlp.veco.token_classification": ["transformers", "torch"], "modelscope.models.nlp.veco.backbone": ["transformers"], "modelscope.models.nlp.veco.configuration": ["transformers"], "modelscope.models.nlp.veco.text_classification": ["transformers"], "modelscope.models.nlp.glm_130b.quantization.functional": ["torch"], "modelscope.models.nlp.glm_130b.quantization.layers": ["SwissArmyTransformer", "torch"], "modelscope.models.nlp.glm_130b.initialize": ["torch", "time", "argparse", "SwissArmyTransformer"], "modelscope.models.nlp.glm_130b.generation.strategies": ["numpy", "SwissArmyTransformer", "torch"], "modelscope.models.nlp.glm_130b.text_generation": ["time", "stat", "re", "functools", "SwissArmyTransformer", "typing", "torch", "sys", "os", "copy", "random"], "modelscope.models.nlp.hf_transformers.backbone": ["transformers"], "modelscope.models.nlp.deberta_v2.fill_mask": ["transformers", "typing", "torch"], "modelscope.models.nlp.deberta_v2.tokenization": ["unicodedata", "transformers", "sentencepiece", "os", "typing"], "modelscope.models.nlp.deberta_v2.backbone": ["collections", "transformers", "typing", "torch"], "modelscope.models.nlp.deberta_v2.configuration": ["transformers"], "modelscope.models.nlp.deberta_v2.tokenization_fast": ["shutil", "transformers", "os", "typing"], "modelscope.models.nlp.bart.text_error_correction": ["os", "typing", "torch"], "modelscope.models.nlp.palm_v2.configuration": ["transformers"], "modelscope.models.nlp.palm_v2.text_generation": ["numpy", "transformers", "codecs", "dataclasses", "math", "subprocess", "json", "typing", "torch", "copy", "os"], "modelscope.models.nlp.palm_v2.dureader_eval": ["numpy", "collections", "re", "math", "json", "argparse", "sys", "zipfile", "copy", "rouge"], "modelscope.models.nlp.space_T_en.text_to_sql": ["text2sql_lgesql", "os", "typing", "torch"], "modelscope.models.nlp.xlm_roberta.backbone": ["math", "packaging", "transformers", "torch"], "modelscope.models.nlp.xlm_roberta.configuration": ["collections", "transformers", "typing"], "modelscope.models.nlp.gpt2.backbone": ["transformers"], "modelscope.models.nlp.codegeex.codegeex_for_code_translation": ["copy", "typing", "torch"], "modelscope.models.nlp.codegeex.codegeex": ["math", "torch"], "modelscope.models.nlp.codegeex.tokenizer": ["transformers", "typing", "torch"], "modelscope.models.nlp.codegeex.codegeex_for_code_generation": ["copy", "typing", "torch"], "modelscope.models.nlp.codegeex.inference": ["typing", "torch"], "modelscope.models.nlp.unite.configuration": ["enum"], "modelscope.models.nlp.unite.translation_evaluation": ["numpy", "transformers", "dataclasses", "math", "packaging", "warnings", "torch", "typing"], "modelscope.models.nlp.space_T_cn.backbone": ["numpy", "__future__", "math", "shutil", "torch", "tarfile", "tempfile", "copy", "os"], "modelscope.models.nlp.space_T_cn.configuration": ["__future__", "logging", "json", "copy"], "modelscope.models.nlp.space_T_cn.table_question_answering": ["numpy", "transformers", "typing", "torch", "os"], "modelscope.models.nlp.megatron_bert.fill_mask": ["transformers", "torch"], "modelscope.models.nlp.megatron_bert.backbone": ["math", "transformers", "torch"], "modelscope.models.nlp.megatron_bert.configuration": ["collections", "transformers", "typing"], "modelscope.models.nlp.llama.backbone": ["transformers"], "modelscope.models.nlp.llama.text_generation": ["transformers", "typing", "torch"], "modelscope.models.nlp.llama.convert_llama_weights_to_hf": ["transformers", "gc", "math", "json", "argparse", "shutil", "torch", "os"], "modelscope.models.nlp.polylm.text_generation": ["collections", "transformers", "typing", "torch"], "modelscope.models.nlp.gpt_moe.backbone": ["transformers", "math", "addict", "typing", "torch", "os"], "modelscope.models.nlp.gpt_moe.configuration": ["transformers", "torch"], "modelscope.models.nlp.gpt_moe.text_generation": ["transformers", "typing"], "modelscope.models.nlp.gpt_moe.distributed_gpt_moe": ["megatron_util", "math", "transformers", "torch"], "modelscope.models.nlp.gpt_moe.tokenizer": ["tokenizers"], "modelscope.models.nlp.gpt_moe.checkpointing": ["megatron_util", "os", "torch"], "modelscope.models.nlp.gpt_moe.moe.layer": ["megatron_util", "typing", "torch"], "modelscope.models.nlp.gpt_moe.moe.sharded_moe": ["tutel", "apex", "megatron_util", "math", "typing", "torch", "scipy"], "modelscope.models.nlp.gpt_moe.moe.utils": ["typing", "torch"], "modelscope.models.nlp.gpt_moe.moe.mappings": ["megatron_util", "torch"], "modelscope.models.nlp.gpt_moe.moe.experts": ["copy", "torch"], "modelscope.models.nlp.space.model.generator": ["numpy", "math", "torch"], "modelscope.models.nlp.space.model.tokenization_space": ["transformers"], "modelscope.models.nlp.space.model.intent_unified_transformer": ["torch"], "modelscope.models.nlp.space.model.unified_transformer": ["numpy", "torch"], "modelscope.models.nlp.space.model.model_base": ["os", "torch"], "modelscope.models.nlp.space.model.gen_unified_transformer": ["torch"], "modelscope.models.nlp.space.dialog_modeling": ["os", "typing"], "modelscope.models.nlp.space.configuration": [], "modelscope.models.nlp.space.dialog_state_tracking": ["transformers", "typing", "torch"], "modelscope.models.nlp.space.modules.functions": ["numpy", "torch"], "modelscope.models.nlp.space.modules.multihead_attention": ["torch"], "modelscope.models.nlp.space.modules.embedder": ["torch"], "modelscope.models.nlp.space.modules.transformer_block": ["torch"], "modelscope.models.nlp.space.modules.feedforward": ["torch"], "modelscope.models.nlp.space.dialog_intent_prediction": ["os", "typing"], "modelscope.models.nlp.bert.fill_mask": [], "modelscope.models.nlp.bert.document_segmentation": ["typing", "torch"], "modelscope.models.nlp.bert.sentence_embedding": ["torch"], "modelscope.models.nlp.bert.siamese_uie": ["torch", "copy"], "modelscope.models.nlp.bert.token_classification": [], "modelscope.models.nlp.bert.backbone": ["math", "packaging", "transformers", "torch"], "modelscope.models.nlp.bert.configuration": ["collections", "transformers", "typing"], "modelscope.models.nlp.bert.word_alignment": ["torch"], "modelscope.models.nlp.bert.text_ranking": [], "modelscope.models.nlp.bert.text_classification": [], "modelscope.models.nlp.bloom.backbone": ["transformers"], "modelscope.models.nlp.bloom.text_generation": ["transformers"], "modelscope.models.nlp.chatglm.tokenization": ["numpy", "transformers", "sentencepiece", "typing", "os"], "modelscope.models.nlp.chatglm.configuration": ["transformers"], "modelscope.models.nlp.chatglm.text_generation": ["transformers", "re", "math", "warnings", "torch", "typing", "sys", "copy", "os"], "modelscope.models.nlp.chatglm.quantization": ["cpm_kernels", "ctypes", "typing", "torch", "base64", "bz2"], "modelscope.models.nlp.fid_T5.text_generation": ["io", "transformers", "os", "torch"], "modelscope.models.nlp.llama2.backbone": [], "modelscope.models.nlp.llama2.text_generation": [], "modelscope.models.nlp.peer.sas_utils": ["nltk", "numpy", "random", "torch"], "modelscope.models.nlp.peer.backbone": ["transformers", "dataclasses", "math", "typing", "torch"], "modelscope.models.nlp.peer.configuration": ["transformers"], "modelscope.models.nlp.peer.text_classification": ["copy", "torch"], "modelscope.models.nlp.canmt.canmt_model": ["numpy", "fairseq", "math", "typing", "torch"], "modelscope.models.nlp.canmt.sequence_generator": ["numpy", "fairseq", "sys", "math", "typing", "torch"], "modelscope.models.nlp.canmt.canmt_translation": ["numpy", "math", "typing", "torch", "os"], "modelscope.models.nlp.qwen.qwen_generation_utils": ["numpy", "transformers", "typing", "torch"], "modelscope.models.nlp.qwen.tokenization": ["__future__", "transformers", "json", "typing", "base64", "unicodedata", "tiktoken", "logging", "os", "io"], "modelscope.models.nlp.qwen.backbone": ["transformers", "math", "typing", "torch", "einops", "importlib", "flash_attn"], "modelscope.models.nlp.qwen.configuration": ["transformers"], "modelscope.models.nlp.qwen.text_generation": ["warnings", "transformers", "typing", "torch"], "modelscope.models.nlp.fid_plug.backbone": ["numpy", "transformers", "dataclasses", "math", "typing", "torch", "copy", "os"], "modelscope.models.nlp.fid_plug.configuration": ["transformers"], "modelscope.models.nlp.fid_plug.text_generation": ["io", "transformers", "os", "torch"], "modelscope.models.nlp.use.transformer": ["math", "torch"], "modelscope.models.nlp.use.user_satisfaction_estimation": ["numpy", "transformers", "typing", "torch", "os"], "modelscope.models.nlp.gpt_neo.backbone": ["transformers"], "modelscope.models.nlp.mglm.model.modeling_glm": ["megatron_util", "torch"], "modelscope.models.nlp.mglm.model.transformer": ["apex", "deepspeed", "megatron_util", "math", "torch"], "modelscope.models.nlp.mglm.model.downstream": ["torch"], "modelscope.models.nlp.mglm.model.prompt": ["random", "torch"], "modelscope.models.nlp.mglm.model.distributed": ["megatron_util", "torch"], "modelscope.models.nlp.mglm.model.modeling_bert": ["__future__", "data_utils", "apex", "megatron_util", "math", "json", "shutil", "torch", "tarfile", "tempfile", "copy", "logging", "os"], "modelscope.models.nlp.mglm.data_utils.datasets": ["numpy", "math", "itertools", "csv", "random", "nltk", "time", "tqdm", "json", "pandas", "torch", "operator", "bisect", "os"], "modelscope.models.nlp.mglm.data_utils.samplers": ["numpy", "sys", "math", "os", "torch"], "modelscope.models.nlp.mglm.data_utils.lazy_loader": ["numpy", "time", "itertools", "mmap", "pickle", "os", "torch"], "modelscope.models.nlp.mglm.data_utils.file_utils": ["functools", "shutil", "urllib", "sys", "hashlib", "requests", "io", "__future__", "tqdm", "json", "botocore", "boto3", "tempfile", "pathlib", "logging", "os"], "modelscope.models.nlp.mglm.data_utils.tokenization": ["nltk", "regex", "collections", "sentencepiece", "torch", "itertools", "os", "csv", "random"], "modelscope.models.nlp.mglm.data_utils.extraction": ["json", "nltk", "os", "glob"], "modelscope.models.nlp.mglm.data_utils.sp_tokenizer": ["os"], "modelscope.models.nlp.mglm.data_utils.wordpiece": ["__future__", "collections", "unicodedata", "logging", "os", "io"], "modelscope.models.nlp.mglm.data_utils.tokenization_gpt2": ["regex", "__future__", "functools", "json", "sys", "logging", "os", "io"], "modelscope.models.nlp.mglm.data_utils.corpora": ["collections", "queue", "tqdm", "json", "torch", "random", "multiprocessing", "os"], "modelscope.models.nlp.mglm.tasks.data_utils": ["numpy", "re", "megatron_util", "pickle", "json", "typing", "torch", "copy"], "modelscope.models.nlp.mglm.tasks.language_model.detokenizer": ["re"], "modelscope.models.nlp.mglm.tasks.language_model.dataset": ["numpy", "utils", "math", "json", "tasks", "torch", "bisect", "itertools"], "modelscope.models.nlp.mglm.tasks.language_model.finetune": ["pretrain_glm", "functools", "math", "torch", "finetune_glm", "tasks", "megatron_util"], "modelscope.models.nlp.mglm.tasks.superglue.pvp": ["numpy", "collections", "utils", "abc", "math", "string", "typing", "tasks", "copy", "random"], "modelscope.models.nlp.mglm.tasks.superglue.dataset": ["numpy", "collections", "abc", "csv", "random", "utils", "re", "tqdm", "json", "pandas", "typing", "torch", "copy", "data_utils", "os", "glob"], "modelscope.models.nlp.mglm.tasks.superglue.finetune": ["collections", "finetune_glm", "tasks"], "modelscope.models.nlp.mglm.tasks.superglue.evaluate": ["__future__", "collections", "functools", "tasks", "string", "typing", "re"], "modelscope.models.nlp.mglm.tasks.eval_utils": ["time", "collections", "utils", "megatron_util", "typing", "torch", "tasks", "datetime", "os", "finetune_glm", "sklearn", "random"], "modelscope.models.nlp.mglm.tasks.seq2seq.dataset": ["numpy", "utils", "tqdm", "json", "tasks", "torch", "os", "data_utils", "random"], "modelscope.models.nlp.mglm.tasks.seq2seq.finetune": ["pretrain_glm", "collections", "functools", "torch", "finetune_glm", "tasks", "megatron_util"], "modelscope.models.nlp.mglm.tasks.seq2seq.evaluate": ["generation_utils", "rouge_score", "datetime", "megatron_util", "random", "string", "torch"], "modelscope.models.nlp.mglm.blocklm_utils": ["numpy", "copy", "megatron_util", "math", "random", "scipy", "torch"], "modelscope.models.nlp.mglm.utils": ["numpy", "time", "megatron_util", "subprocess", "json", "torch", "os", "random"], "modelscope.models.nlp.mglm.arguments": ["deepspeed", "json", "argparse", "torch", "os"], "modelscope.models.nlp.mglm.run_test": ["sys", "test"], "modelscope.models.nlp.mglm.generation_utils": ["abc", "collections", "typing", "torch"], "modelscope.models.nlp.mglm.train_utils": ["deepspeed", "megatron_util", "apex", "torch"], "modelscope.models.nlp.mglm.test.test_block": ["numpy", "blocklm_utils", "random", "argparse"], "modelscope.models.nlp.mglm.test.test_rel_shift": ["numpy", "learning_rates", "matplotlib", "torch"], "modelscope.models.nlp.mglm.configure_data": ["numpy", "megatron_util", "torch", "bisect", "itertools", "os", "copy", "random"], "modelscope.models.nlp.mglm.process_grid": ["json", "sys", "os", "statistics", "glob"], "modelscope.models.nlp.mglm.mglm_for_text_summarization": ["numpy", "megatron_util", "typing", "torch", "random", "os"], "modelscope.models.nlp.lstm.token_classification": [], "modelscope.models.nlp.lstm.backbone": ["torch"], "modelscope.models.nlp.dgds.document_grounded_dialog_retrieval": ["os", "typing", "torch"], "modelscope.models.nlp.dgds.backbone": ["__future__", "transformers", "os", "torch"], "modelscope.models.nlp.dgds.document_grounded_dialog_rerank": ["os", "typing", "torch"], "modelscope.models.nlp.dgds.document_grounded_dialog_generate": ["os", "typing", "torch"], "modelscope.models.nlp.plug.generator": ["torch"], "modelscope.models.nlp.plug.AnnealingLR": ["math", "torch"], "modelscope.models.nlp.plug.backbone": ["__future__", "megatron_util", "math", "torch", "logging"], "modelscope.models.nlp.plug.configuration": ["json", "transformers", "copy"], "modelscope.models.nlp.plug.distributed_plug": ["megatron_util", "typing", "torch"], "modelscope.models.nlp.structbert.fill_mask": ["transformers", "torch"], "modelscope.models.nlp.structbert.faq_question_answering": ["collections", "math", "typing", "torch", "os"], "modelscope.models.nlp.structbert.token_classification": ["torch"], "modelscope.models.nlp.structbert.backbone": ["transformers", "dataclasses", "math", "packaging", "typing", "torch"], "modelscope.models.nlp.structbert.configuration": ["transformers"], "modelscope.models.nlp.structbert.adv_utils": ["torch"], "modelscope.models.nlp.structbert.text_classification": ["torch"], "modelscope.models.nlp.plug_mental.backbone": ["transformers", "dataclasses", "math", "packaging", "typing", "torch"], "modelscope.models.nlp.plug_mental.configuration": ["transformers"], "modelscope.models.nlp.plug_mental.adv_utils": ["torch"], "modelscope.models.nlp.plug_mental.text_classification": ["torch"], "modelscope.models.nlp.gpt3.distributed_gpt3": ["collections", "transformers", "megatron_util", "math", "typing", "torch", "os"], "modelscope.models.nlp.gpt3.backbone": ["transformers", "math", "addict", "typing", "torch", "os"], "modelscope.models.nlp.gpt3.configuration": ["transformers", "torch"], "modelscope.models.nlp.gpt3.text_generation": ["collections", "transformers", "typing", "torch"], "modelscope.models.nlp.gpt3.tokenizer": ["typing", "tokenizers"], "modelscope.models.nlp.chatglm2.tokenization": ["transformers", "os", "typing", "sentencepiece"], "modelscope.models.nlp.chatglm2.configuration": ["transformers"], "modelscope.models.nlp.chatglm2.text_generation": ["transformers", "math", "typing", "warnings", "torch", "sys", "copy"], "modelscope.models.nlp.chatglm2.quantization": ["cpm_kernels", "ctypes", "typing", "torch", "base64", "bz2"], "modelscope.models.nlp.csanmt.translation": ["tensorflow", "math", "collections", "typing"], "modelscope.models.nlp.task_models.fill_mask": ["numpy", "typing", "torch"], "modelscope.models.nlp.task_models.information_extraction": ["numpy", "typing"], "modelscope.models.nlp.task_models.token_classification": ["typing", "torch"], "modelscope.models.nlp.task_models.machine_reading_comprehension": ["transformers", "dataclasses", "typing", "torch", "os"], "modelscope.models.nlp.task_models.text_generation": ["numpy", "transformers", "typing", "torch"], "modelscope.models.nlp.task_models.text_ranking": ["numpy", "typing"], "modelscope.models.nlp.task_models.text_classification": ["numpy", "typing"], "modelscope.models.nlp.task_models.task_model": ["collections", "re", "abc", "typing", "torch", "os"], "modelscope.models.nlp.task_models.feature_extraction": ["numpy", "typing"], "modelscope.models.nlp.T5.text2text_generation": ["transformers", "warnings", "typing", "torch", "copy"], "modelscope.models.nlp.T5.backbone": ["transformers", "math", "typing", "warnings", "torch", "copy", "os"], "modelscope.models.nlp.T5.configuration": ["transformers", "typing"], "modelscope.models.science.unifold.msa.parsers": ["collections", "itertools", "dataclasses", "string", "typing", "re"], "modelscope.models.science.unifold.msa.mmcif": ["collections", "absl", "Bio", "dataclasses", "functools", "io", "typing"], "modelscope.models.science.unifold.msa.utils": ["typing", "json", "os", "absl"], "modelscope.models.science.unifold.msa.pipeline": ["numpy", "typing", "os", "absl"], "modelscope.models.science.unifold.msa.tools.hhblits": ["absl", "subprocess", "typing", "os", "glob"], "modelscope.models.science.unifold.msa.tools.hmmbuild": ["re", "subprocess", "os", "absl"], "modelscope.models.science.unifold.msa.tools.hhsearch": ["absl", "subprocess", "typing", "os", "glob"], "modelscope.models.science.unifold.msa.tools.utils": ["contextlib", "time", "tempfile", "absl", "typing", "shutil"], "modelscope.models.science.unifold.msa.tools.kalign": ["subprocess", "typing", "os", "absl"], "modelscope.models.science.unifold.msa.tools.jackhmmer": ["absl", "subprocess", "typing", "glob", "urllib", "os", "concurrent"], "modelscope.models.science.unifold.msa.tools.hmmsearch": ["subprocess", "typing", "os", "absl"], "modelscope.models.science.unifold.msa.msa_identifiers": ["dataclasses", "typing", "re"], "modelscope.models.science.unifold.msa.templates": ["numpy", "absl", "abc", "functools", "dataclasses", "re", "typing", "datetime", "os", "glob"], "modelscope.models.science.unifold.dataset": ["numpy", "ml_collections", "json", "typing", "torch", "copy", "logging", "os", "unicore"], "modelscope.models.science.unifold.modules.structure_module": ["math", "unicore", "typing", "torch"], "modelscope.models.science.unifold.modules.auxillary_heads": ["unicore", "typing", "torch"], "modelscope.models.science.unifold.modules.featurization": ["unicore", "typing", "torch"], "modelscope.models.science.unifold.modules.triangle_multiplication": ["functools", "unicore", "typing", "torch"], "modelscope.models.science.unifold.modules.alphafold": ["unicore", "torch"], "modelscope.models.science.unifold.modules.confidence": ["typing", "torch"], "modelscope.models.science.unifold.modules.evoformer": ["functools", "unicore", "typing", "torch"], "modelscope.models.science.unifold.modules.template": ["functools", "math", "typing", "torch", "unicore"], "modelscope.models.science.unifold.modules.attentions": ["functools", "unicore", "typing", "torch"], "modelscope.models.science.unifold.modules.embedders": ["unicore", "typing", "torch"], "modelscope.models.science.unifold.modules.common": ["functools", "unicore", "typing", "torch"], "modelscope.models.science.unifold.modules.frame": ["numpy", "__future__", "typing", "torch"], "modelscope.models.science.unifold.config": ["ml_collections", "typing", "copy"], "modelscope.models.science.unifold.model": ["os", "argparse", "typing", "torch"], "modelscope.models.science.unifold.data.residue_constants": ["numpy", "collections", "functools", "os", "unicore", "typing"], "modelscope.models.science.unifold.data.msa_pairing": ["numpy", "collections", "pandas", "typing", "scipy"], "modelscope.models.science.unifold.data.data_ops": ["numpy", "functools", "typing", "operator", "torch", "itertools", "unicore"], "modelscope.models.science.unifold.data.utils": ["numpy", "gzip", "functools", "pickle", "json", "typing", "copy", "scipy"], "modelscope.models.science.unifold.data.protein": ["numpy", "dataclasses", "typing", "Bio", "io"], "modelscope.models.science.unifold.data.process_multimer": ["numpy", "collections", "typing"], "modelscope.models.science.unifold.data.process": ["numpy", "typing", "torch"], "modelscope.models.audio.punc.generic_punctuation": ["os", "typing"], "modelscope.models.audio.itn.generic_inverse_text_processing": ["os", "typing"], "modelscope.models.audio.aec.layers.affine_transform": ["numpy", "torch"], "modelscope.models.audio.aec.layers.layer_base": ["abc", "torch", "numpy", "re"], "modelscope.models.audio.aec.layers.activations": ["torch"], "modelscope.models.audio.aec.layers.uni_deep_fsmn": ["numpy", "torch"], "modelscope.models.audio.aec.layers.deep_fsmn": ["numpy", "torch"], "modelscope.models.audio.aec.network.loss": ["torch"], "modelscope.models.audio.aec.network.se_net": ["torch"], "modelscope.models.audio.aec.network.modulation_loss": ["math", "torch", "torchaudio"], "modelscope.models.audio.ans.layers.affine_transform": ["torch"], "modelscope.models.audio.ans.layers.layer_base": ["abc", "numpy", "six", "torch"], "modelscope.models.audio.ans.layers.activations": ["torch"], "modelscope.models.audio.ans.layers.uni_deep_fsmn": ["numpy", "torch"], "modelscope.models.audio.ans.denoise_net": ["torch"], "modelscope.models.audio.ans.unet": ["torch"], "modelscope.models.audio.ans.se_module_complex": ["torch"], "modelscope.models.audio.ans.conv_stft": ["numpy", "scipy", "torch"], "modelscope.models.audio.ans.complex_nn": ["torch"], "modelscope.models.audio.ans.frcrn": ["os", "typing", "torch"], "modelscope.models.audio.separation.m2.fsmn": ["torch"], "modelscope.models.audio.separation.m2.mossformer_block": ["rotary_embedding_torch", "einops", "torch"], "modelscope.models.audio.separation.m2.conv_module": ["torch"], "modelscope.models.audio.separation.m2.mossformer": ["os", "torch"], "modelscope.models.audio.separation.m2.layer_norm": ["__future__", "torch"], "modelscope.models.audio.separation.mossformer_block": ["torch"], "modelscope.models.audio.separation.mossformer": ["copy", "os", "typing", "torch"], "modelscope.models.audio.separation.layer_norm": ["__future__", "torch"], "modelscope.models.audio.separation.mossformer_conv_module": ["torch"], "modelscope.models.audio.sv.rdino": ["math", "typing", "torch", "os", "torchaudio"], "modelscope.models.audio.sv.speaker_diarization_semantic_speaker_turn_detection": ["torch"], "modelscope.models.audio.sv.DTDNN": ["numpy", "collections", "typing", "torch", "os", "torchaudio"], "modelscope.models.audio.sv.ERes2Net": ["math", "typing", "torch", "os", "torchaudio"], "modelscope.models.audio.sv.ERes2Net_aug": ["math", "typing", "torch", "os", "torchaudio"], "modelscope.models.audio.sv.cluster_backend": ["numpy", "umap", "typing", "sklearn", "hdbscan", "scipy"], "modelscope.models.audio.sv.ecapa_tdnn": ["numpy", "math", "typing", "torch", "os", "torchaudio"], "modelscope.models.audio.sv.generic_speaker_verification": ["os", "typing"], "modelscope.models.audio.sv.speaker_change_locator": ["numpy", "collections", "typing", "torch", "os", "torchaudio"], "modelscope.models.audio.sv.speaker_diarization_dialogue_detection": ["torch"], "modelscope.models.audio.sv.pooling_layers": ["torch"], "modelscope.models.audio.sv.lanuage_recognition_eres2net": ["numpy", "typing", "torch", "os", "torchaudio"], "modelscope.models.audio.sv.fusion": ["torch"], "modelscope.models.audio.sv.lanuage_recognition_model": ["numpy", "typing", "torch", "os", "torchaudio"], "modelscope.models.audio.sv.DTDNN_layers": ["torch"], "modelscope.models.audio.tts.sambert_hifi": ["wave", "numpy", "__future__", "yaml", "json", "shutil", "matplotlib", "datetime", "zipfile", "os"], "modelscope.models.audio.tts.voice": ["numpy", "time", "collections", "kantts", "yaml", "pickle", "json", "torch", "threading", "os"], "modelscope.models.audio.kws.farfield.fsmn": ["numpy", "torch"], "modelscope.models.audio.kws.farfield.model_def": ["enum", "math", "struct"], "modelscope.models.audio.kws.farfield.fsmn_sele_v3": ["torch"], "modelscope.models.audio.kws.farfield.model": ["tempfile", "os", "typing"], "modelscope.models.audio.kws.farfield.fsmn_sele_v2": ["torch"], "modelscope.models.audio.kws.nearfield.fsmn": ["numpy", "typing", "torch"], "modelscope.models.audio.kws.nearfield.cmvn": ["re", "numpy", "torch"], "modelscope.models.audio.kws.nearfield.model": ["typing", "torch", "tempfile", "sys", "os"], "modelscope.models.audio.kws.generic_key_word_spotting": ["os", "typing"], "modelscope.models.audio.asr.generic_automatic_speech_recognition": ["os", "typing"], "modelscope.models.audio.asr.wenet_automatic_speech_recognition": ["os", "json", "wenetruntime", "typing"], "modelscope.models.cv.image_colorization.unet.utils": ["functools", "enum", "torch"], "modelscope.models.cv.image_colorization.unet.unet": ["numpy", "torch"], "modelscope.models.cv.image_colorization.ddcolor.loss": ["torch"], "modelscope.models.cv.image_colorization.ddcolor.utils.vgg": ["collections", "os", "torchvision", "torch"], "modelscope.models.cv.image_colorization.ddcolor.utils.position_encoding": ["math", "torch"], "modelscope.models.cv.image_colorization.ddcolor.utils.transformer_utils": ["typing", "torch"], "modelscope.models.cv.image_colorization.ddcolor.utils.convnext": ["timm", "torch"], "modelscope.models.cv.image_colorization.ddcolor.utils.unet": ["enum", "collections", "torch"], "modelscope.models.cv.image_colorization.ddcolor.ddcolor_for_image_colorization": ["numpy", "typing", "torch", "copy", "os"], "modelscope.models.cv.image_colorization.ddcolor.ddcolor": ["torch"], "modelscope.models.cv.face_detection.peppa_pig_face.facer": ["numpy", "cv2"], "modelscope.models.cv.face_detection.peppa_pig_face.face_landmark": ["numpy", "tensorflow", "cv2"], "modelscope.models.cv.face_detection.peppa_pig_face.face_detector": ["numpy", "tensorflow", "cv2"], "modelscope.models.cv.face_detection.peppa_pig_face.LK.lk": ["numpy"], "modelscope.models.cv.face_detection.retinaface.models.net": ["time", "torchvision", "torch"], "modelscope.models.cv.face_detection.retinaface.models.retinaface": ["collections", "torchvision", "torch"], "modelscope.models.cv.face_detection.retinaface.utils": ["numpy", "math", "itertools", "torch"], "modelscope.models.cv.face_detection.retinaface.detection": ["numpy", "cv2", "torch"], "modelscope.models.cv.face_detection.ulfd_slim.vision.transforms": ["numpy", "torch", "types", "cv2"], "modelscope.models.cv.face_detection.ulfd_slim.vision.box_utils": ["math", "torch"], "modelscope.models.cv.face_detection.ulfd_slim.vision.mb_tiny": ["torch"], "modelscope.models.cv.face_detection.ulfd_slim.vision.ssd.ssd": ["numpy", "collections", "typing", "torch"], "modelscope.models.cv.face_detection.ulfd_slim.vision.ssd.data_preprocessing": [], "modelscope.models.cv.face_detection.ulfd_slim.vision.ssd.mb_tiny_fd": ["torch"], "modelscope.models.cv.face_detection.ulfd_slim.vision.ssd.predictor": ["torch"], "modelscope.models.cv.face_detection.ulfd_slim.vision.ssd.fd_config": ["numpy"], "modelscope.models.cv.face_detection.ulfd_slim.detection": ["numpy", "cv2", "os", "torch"], "modelscope.models.cv.face_detection.scrfd.preprocessor": ["PIL", "numpy", "typing"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.core.bbox.transforms": ["numpy", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.core.post_processing.bbox_nms": ["torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.detectors.base": ["numpy", "collections", "mmcv", "abc", "mmdet", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.detectors.tinymog": ["mmdet", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.detectors.scrfd": ["mmdet", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.detectors.single_stage": ["mmdet", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.dense_heads.scrfd_head": ["mmcv", "numpy", "mmdet", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.backbones.master_net": ["mmdet", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.backbones.resnet": ["mmcv", "mmdet", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.models.backbones.mobilenet": ["mmcv", "mmdet", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.transforms": ["mmcv", "mmdet", "numpy"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.formating": ["mmcv", "numpy", "mmdet", "torch"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.loading": ["numpy", "mmdet", "os", "pycocotools"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.pipelines.auto_augment": ["numpy", "copy", "mmcv", "mmdet", "cv2"], "modelscope.models.cv.face_detection.scrfd.mmdet_patch.datasets.retinaface": ["numpy", "mmdet"], "modelscope.models.cv.face_detection.scrfd.tinymog_detect": ["copy", "os", "typing", "torch"], "modelscope.models.cv.face_detection.scrfd.scrfd_detect": ["numpy", "typing", "torch", "copy", "os"], "modelscope.models.cv.face_detection.scrfd.damofd_detect": ["copy", "os", "typing", "torch"], "modelscope.models.cv.face_detection.mtcnn.models.first_stage": ["PIL", "numpy", "math", "torch"], "modelscope.models.cv.face_detection.mtcnn.models.box_utils": ["PIL", "numpy"], "modelscope.models.cv.face_detection.mtcnn.models.detector": ["PIL", "numpy", "os", "torch"], "modelscope.models.cv.face_detection.mtcnn.models.get_nets": ["numpy", "collections", "torch"], "modelscope.models.cv.face_detection.mogface.models.detectors": ["numpy", "cv2", "os", "torch"], "modelscope.models.cv.face_detection.mogface.models.mogface": ["torch"], "modelscope.models.cv.face_detection.mogface.models.utils": ["numpy", "math", "itertools", "torch"], "modelscope.models.cv.face_detection.mogface.models.resnet": ["torch"], "modelscope.models.cv.face_detection.mogface.models.mogprednet": ["math", "torch"], "modelscope.models.cv.product_segmentation.seg_infer": ["PIL", "numpy", "cv2", "torch"], "modelscope.models.cv.product_segmentation.net": ["torch"], "modelscope.models.cv.video_summarization.kts.cpd_auto": ["numpy"], "modelscope.models.cv.video_summarization.kts.cpd_nonlin": ["numpy"], "modelscope.models.cv.video_summarization.base_model": ["numpy", "cv2", "torch"], "modelscope.models.cv.video_summarization.pgl_sum": ["math", "torch"], "modelscope.models.cv.video_summarization.summarizer": ["numpy", "os", "typing", "torch"], "modelscope.models.cv.image_skychange.preprocessor": ["numpy", "torchvision", "json", "numbers", "typing", "cv2", "pdb", "torch"], "modelscope.models.cv.image_skychange.ptsemseg.hrnet_super_and_ocr": ["numpy", "__future__", "torch"], "modelscope.models.cv.image_skychange.ptsemseg.BlockModules": ["torch"], "modelscope.models.cv.image_skychange.ptsemseg.unet": ["torch"], "modelscope.models.cv.image_skychange.ptsemseg.hrnet_backnone": ["numpy", "logging", "os", "torch"], "modelscope.models.cv.image_skychange.skychange": ["numpy", "collections", "torchvision", "json", "numbers", "cv2", "pdb", "torch", "PIL", "os"], "modelscope.models.cv.image_skychange.skychange_model": ["collections", "time", "math", "json", "typing", "cv2", "pdb", "torch", "os"], "modelscope.models.cv.cmdssl_video_embedding.c3d": ["torch"], "modelscope.models.cv.cmdssl_video_embedding.resnet3d": ["torch"], "modelscope.models.cv.cmdssl_video_embedding.resnet2p1d": ["torch"], "modelscope.models.cv.tinynas_detection.damo.augmentations.scale_aware_aug": ["copy"], "modelscope.models.cv.tinynas_detection.damo.augmentations.box_level_augs.geometric_augs": ["copy", "random", "torchvision", "torch"], "modelscope.models.cv.tinynas_detection.damo.augmentations.box_level_augs.gaussian_maps": ["math", "torch"], "modelscope.models.cv.tinynas_detection.damo.augmentations.box_level_augs.color_augs": ["random", "torch"], "modelscope.models.cv.tinynas_detection.damo.augmentations.box_level_augs.box_level_augs": ["numpy", "random"], "modelscope.models.cv.tinynas_detection.damo.base_models.heads.gfocal_v2_tiny": ["functools", "numpy", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.heads.zero_head": ["torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.core.ota_assigner": ["warnings", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.core.weight_init": ["numpy", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.core.utils": ["functools", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.core.base_ops": ["math", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.core.neck_ops": ["numpy", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.core.repvgg_block": ["numpy", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.core.ops": ["numpy", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.losses.gfocal_loss": ["functools", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.losses.distill_loss": ["torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.backbones.darknet": ["torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.backbones.tinynas_csp": ["torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.backbones.tinynas_res": ["torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.necks.giraffe_fpn": ["numpy", "collections", "functools", "math", "timm", "typing", "torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.necks.giraffe_fpn_btn": ["torch"], "modelscope.models.cv.tinynas_detection.damo.base_models.necks.giraffe_config": ["collections", "networkx"], "modelscope.models.cv.tinynas_detection.damo.detectors.detector": ["torch"], "modelscope.models.cv.tinynas_detection.damo.structures.bounding_box": ["torch"], "modelscope.models.cv.tinynas_detection.damo.structures.image_list": ["__future__", "torch"], "modelscope.models.cv.tinynas_detection.damo.structures.boxlist_ops": ["torch"], "modelscope.models.cv.tinynas_detection.damo.utils.model_utils": ["time", "copy", "math", "thop", "torch"], "modelscope.models.cv.tinynas_detection.damo.utils.boxes": ["numpy", "torchvision", "torch"], "modelscope.models.cv.tinynas_detection.damo.utils.scheduler": ["math"], "modelscope.models.cv.tinynas_detection.damo.apis.detector_evaluater": ["os", "torch"], "modelscope.models.cv.tinynas_detection.damo.apis.detector_inference": ["tqdm", "os", "torch"], "modelscope.models.cv.tinynas_detection.detector": ["pickle", "os", "torchvision", "torch"], "modelscope.models.cv.tinynas_detection.utils": ["tempfile", "sys", "importlib", "easydict", "os", "shutil"], "modelscope.models.cv.tinynas_detection.tinynas_detector": [], "modelscope.models.cv.tinynas_detection.tinynas_damoyolo": [], "modelscope.models.cv.face_emotion.face_alignment.face_align": ["numpy", "cv2", "PIL", "sys", "os"], "modelscope.models.cv.face_emotion.face_alignment.face": ["numpy", "os", "tensorflow", "cv2"], "modelscope.models.cv.face_emotion.emotion_model": ["os", "sys", "torch"], "modelscope.models.cv.face_emotion.efficient.utils": ["collections", "functools", "math", "torch", "re"], "modelscope.models.cv.face_emotion.efficient.model": ["torch"], "modelscope.models.cv.face_emotion.emotion_infer": ["PIL", "torchvision", "torch"], "modelscope.models.cv.motion_generation.modules.gaussian_diffusion": ["enum", "numpy", "copy", "math", "torch"], "modelscope.models.cv.motion_generation.modules.respace": ["numpy", "torch"], "modelscope.models.cv.motion_generation.modules.cfg_sampler": ["torch", "copy"], "modelscope.models.cv.motion_generation.modules.smpl": ["numpy", "smplx", "os", "contextlib", "torch"], "modelscope.models.cv.motion_generation.modules.mdm": ["clip", "numpy", "torch"], "modelscope.models.cv.motion_generation.modules.rotation2xyz": ["torch"], "modelscope.models.cv.motion_generation.model": [], "modelscope.models.cv.image_quality_assessment_mos.heads.simple_head": ["torch"], "modelscope.models.cv.image_quality_assessment_mos.image_quality_assessment_mos": ["os", "typing", "torch"], "modelscope.models.cv.image_quality_assessment_mos.censeo_ivqa_model": ["torch"], "modelscope.models.cv.image_quality_assessment_mos.backbones.resnet": ["os", "torch"], "modelscope.models.cv.image_quality_assessment_degradation.degradation_model": ["numpy", "time", "collections", "torchvision", "cv2", "json", "torch"], "modelscope.models.cv.image_quality_assessment_degradation.image_quality_assessment_degradation": ["os", "typing", "torch"], "modelscope.models.cv.pedestrian_attribute_recognition.model": ["numpy", "os", "torchvision", "torch"], "modelscope.models.cv.image_depth_estimation.newcrfs_model": ["numpy", "os", "torch"], "modelscope.models.cv.image_depth_estimation.networks.uper_crf_head": ["mmcv", "torch"], "modelscope.models.cv.image_depth_estimation.networks.swin_transformer": ["numpy", "timm", "torch"], "modelscope.models.cv.image_depth_estimation.networks.newcrf_layers": ["numpy", "timm", "torch"], "modelscope.models.cv.image_depth_estimation.networks.newcrf_utils": ["collections", "torchvision", "importlib", "pkgutil", "os", "warnings", "torch"], "modelscope.models.cv.image_depth_estimation.networks.newcrf_depth": ["torch"], "modelscope.models.cv.face_generation.op.conv2d_gradfix": ["contextlib", "warnings", "torch"], "modelscope.models.cv.face_generation.op.fused_act": ["os", "torch"], "modelscope.models.cv.face_generation.op.upfirdn2d": ["collections", "os", "torch"], "modelscope.models.cv.face_generation.stylegan2": ["functools", "math", "operator", "torch", "random"], "modelscope.models.cv.panorama_depth_estimation.unifuse_model": ["numpy", "os", "torchvision", "torch"], "modelscope.models.cv.panorama_depth_estimation.networks.unifuse": ["numpy", "__future__", "collections", "torch"], "modelscope.models.cv.panorama_depth_estimation.networks.equi": ["numpy", "__future__", "collections", "torch"], "modelscope.models.cv.panorama_depth_estimation.networks.resnet": ["torch"], "modelscope.models.cv.panorama_depth_estimation.networks.mobilenet": ["torch"], "modelscope.models.cv.panorama_depth_estimation.networks.util": ["numpy", "scipy", "cv2"], "modelscope.models.cv.panorama_depth_estimation.networks.layers": ["numpy", "torch"], "modelscope.models.cv.tinynas_classfication.global_utils": [], "modelscope.models.cv.tinynas_classfication.plain_net_utils": ["torch"], "modelscope.models.cv.tinynas_classfication.super_res_idwexkx": ["uuid", "torch"], "modelscope.models.cv.tinynas_classfication.model_zoo": [], "modelscope.models.cv.tinynas_classfication.super_blocks": ["uuid", "torch"], "modelscope.models.cv.tinynas_classfication.super_res_kxkx": ["uuid", "torch"], "modelscope.models.cv.tinynas_classfication.master_net": ["torch"], "modelscope.models.cv.tinynas_classfication.super_res_k1kxk1": ["uuid", "torch"], "modelscope.models.cv.tinynas_classfication.basic_blocks": ["numpy", "uuid", "torch"], "modelscope.models.cv.bad_image_detecting.bad_image_detecting": ["numpy", "torchvision", "typing", "torch", "os"], "modelscope.models.cv.skin_retouching.retinaface.predict_single": ["numpy", "torchvision", "typing", "torch", "albumentations"], "modelscope.models.cv.skin_retouching.retinaface.prior_box": ["math", "itertools", "torch"], "modelscope.models.cv.skin_retouching.retinaface.box_utils": ["numpy", "typing", "torch"], "modelscope.models.cv.skin_retouching.retinaface.net": ["typing", "torch"], "modelscope.models.cv.skin_retouching.retinaface.utils": ["numpy", "pathlib", "re", "cv2", "typing", "torch"], "modelscope.models.cv.skin_retouching.retinaface.network": ["torchvision", "typing", "torch"], "modelscope.models.cv.skin_retouching.utils": ["numpy", "time", "einops", "cv2", "typing", "torch"], "modelscope.models.cv.skin_retouching.inpainting_model.inpainting_unet": ["torch"], "modelscope.models.cv.skin_retouching.inpainting_model.gconv": ["torch"], "modelscope.models.cv.skin_retouching.detection_model.detection_unet_in": ["torch"], "modelscope.models.cv.skin_retouching.detection_model.detection_module": ["torch"], "modelscope.models.cv.skin_retouching.unet_deploy": ["warnings", "torch"], "modelscope.models.cv.skin_retouching.weights_init": ["torch"], "modelscope.models.cv.animal_recognition.resnet": ["math", "torch"], "modelscope.models.cv.animal_recognition.splat": ["torch"], "modelscope.models.cv.video_super_resolution.common": ["torch"], "modelscope.models.cv.video_super_resolution.real_basicvsr_for_video_super_resolution": ["os", "typing", "torch"], "modelscope.models.cv.video_super_resolution.msrresnet_lite_model": ["functools", "os", "typing", "torch"], "modelscope.models.cv.video_super_resolution.basicvsr_net": ["torch"], "modelscope.models.cv.video_super_resolution.real_basicvsr_net": ["torch"], "modelscope.models.cv.virual_tryon.sdafnet": ["numpy", "random", "torch"], "modelscope.models.cv.ocr_recognition.preprocessor": ["numpy", "cv2", "PIL", "torch", "os"], "modelscope.models.cv.ocr_recognition.modules.ConvNextViT.timm_tinyc": ["collections", "itertools", "copy", "functools", "math", "logging", "torch"], "modelscope.models.cv.ocr_recognition.modules.ConvNextViT.main_model": ["torch"], "modelscope.models.cv.ocr_recognition.modules.ConvNextViT.convnext": ["torch"], "modelscope.models.cv.ocr_recognition.modules.ConvNextViT.vitstr": ["__future__", "functools", "torch", "copy", "logging"], "modelscope.models.cv.ocr_recognition.modules.CRNN.main_model": ["torch"], "modelscope.models.cv.ocr_recognition.modules.LightweightEdge.main_model": ["collections", "torch"], "modelscope.models.cv.ocr_recognition.modules.LightweightEdge.nas_block.mix_ops": ["numpy", "torch"], "modelscope.models.cv.ocr_recognition.modules.LightweightEdge.nas_block.proxyless": ["numpy", "re", "queue", "torch", "sys"], "modelscope.models.cv.ocr_recognition.modules.LightweightEdge.nas_block.layers": ["numpy", "collections", "torch"], "modelscope.models.cv.ocr_recognition.model": ["os", "torch"], "modelscope.models.cv.image_debanding.rrdb.rrdb_image_debanding": ["os", "typing", "torch"], "modelscope.models.cv.crowd_counting.hrnet_aspp_relu": ["numpy", "functools", "torch", "logging", "os"], "modelscope.models.cv.crowd_counting.cc_model": ["os", "typing", "torch"], "modelscope.models.cv.video_frame_interpolation.VFINet_arch": ["torch"], "modelscope.models.cv.video_frame_interpolation.VFINet_for_video_frame_interpolation": ["copy", "os", "typing", "torch"], "modelscope.models.cv.video_frame_interpolation.flow_model.extractor": ["torch"], "modelscope.models.cv.video_frame_interpolation.flow_model.update": ["torch"], "modelscope.models.cv.video_frame_interpolation.flow_model.raft": ["numpy", "torch"], "modelscope.models.cv.video_frame_interpolation.flow_model.corr": ["torch"], "modelscope.models.cv.video_frame_interpolation.utils.utils": ["numpy", "scipy", "torch"], "modelscope.models.cv.video_frame_interpolation.utils.scene_change_detection": ["numpy", "torch"], "modelscope.models.cv.video_frame_interpolation.interp_model.refinenet_arch": ["numpy", "torch"], "modelscope.models.cv.video_frame_interpolation.interp_model.UNet": ["torch"], "modelscope.models.cv.video_frame_interpolation.interp_model.IFNet_swin": ["numpy", "timm", "torch"], "modelscope.models.cv.video_frame_interpolation.interp_model.flow_reversal": ["torch"], "modelscope.models.cv.video_frame_interpolation.interp_model.transformer_layers": ["sys", "functools", "math", "timm", "torch"], "modelscope.models.cv.face_attribute_recognition.fair_face.face_attribute_recognition": ["numpy", "torchvision", "cv2", "PIL", "torch", "os"], "modelscope.models.cv.image_to_image_generation.models.autoencoder": ["math", "torch"], "modelscope.models.cv.image_to_image_generation.models.clip": ["math", "torch"], "modelscope.models.cv.image_to_image_generation.ops.diffusion": ["math", "torch"], "modelscope.models.cv.image_to_image_generation.ops.losses": ["math", "torch"], "modelscope.models.cv.image_to_image_generation.model": ["math", "torch"], "modelscope.models.cv.image_to_image_generation.data.transforms": ["PIL", "math", "random", "torchvision"], "modelscope.models.cv.pointcloud_sceneflow_estimation.rcp_model": ["numpy", "os", "torch"], "modelscope.models.cv.pointcloud_sceneflow_estimation.sf_rcp": ["torch"], "modelscope.models.cv.pointcloud_sceneflow_estimation.pointnet2_utils": ["pointnet2_cuda", "typing", "torch"], "modelscope.models.cv.pointcloud_sceneflow_estimation.common": ["math", "torch"], "modelscope.models.cv.image_panoptic_segmentation.panseg_model": ["os", "torch"], "modelscope.models.cv.image_try_on.generator": ["functools", "os", "torchvision", "torch"], "modelscope.models.cv.image_try_on.warping": ["numpy", "collections", "math", "cv2", "torch"], "modelscope.models.cv.image_try_on.try_on_infer": ["numpy", "torchvision", "yaml", "argparse", "torch", "cv2", "PIL", "os"], "modelscope.models.cv.image_try_on.landmark": ["logging", "os", "torch"], "modelscope.models.cv.image_paintbyexample.model": ["typing", "torch", "omegaconf", "paint_ldm", "os"], "modelscope.models.cv.face_recognition.align_face": ["skimage", "numpy", "cv2"], "modelscope.models.cv.face_recognition.torchkit.rts_backbone": ["math", "collections", "os", "torch"], "modelscope.models.cv.face_recognition.torchkit.backbone.model_resnet": ["torch"], "modelscope.models.cv.face_recognition.torchkit.backbone.model_irse": ["collections", "torch"], "modelscope.models.cv.face_recognition.torchkit.backbone.common": ["torch"], "modelscope.models.cv.face_recognition.torchkit.backbone.facemask_backbone": ["collections", "torch"], "modelscope.models.cv.face_recognition.torchkit.backbone.arcface_backbone": ["torch"], "modelscope.models.cv.image_instance_segmentation.maskdino_model": ["os", "typing", "torch"], "modelscope.models.cv.image_instance_segmentation.maskdino_swin": ["os", "torch"], "modelscope.models.cv.image_instance_segmentation.fastinst_model": ["os", "typing", "torch"], "modelscope.models.cv.image_instance_segmentation.maskdino.ms_deform_attn": ["__future__", "mmcv", "math", "warnings", "torch"], "modelscope.models.cv.image_instance_segmentation.maskdino.position_encoding": ["math", "torch"], "modelscope.models.cv.image_instance_segmentation.maskdino.maskdino_encoder": ["numpy", "typing", "torch"], "modelscope.models.cv.image_instance_segmentation.maskdino.utils": ["math", "copy", "torch"], "modelscope.models.cv.image_instance_segmentation.maskdino.maskdino_decoder": ["torch"], "modelscope.models.cv.image_instance_segmentation.maskdino.dino_decoder": ["typing", "torch"], "modelscope.models.cv.image_instance_segmentation.model": ["os", "typing", "torch"], "modelscope.models.cv.image_instance_segmentation.fastinst.fastinst_encoder": ["logging", "typing", "torch"], "modelscope.models.cv.image_instance_segmentation.fastinst.fastinst_decoder": ["math", "torch"], "modelscope.models.cv.image_instance_segmentation.postprocess_utils": ["numpy", "torch", "cv2", "pycocotools", "itertools"], "modelscope.models.cv.image_instance_segmentation.cascade_mask_rcnn_swin": ["collections", "os", "torch"], "modelscope.models.cv.image_instance_segmentation.backbones.swin_transformer": ["numpy", "timm", "torch"], "modelscope.models.cv.image_instance_segmentation.backbones.resnet": ["torch"], "modelscope.models.cv.image_instance_segmentation.datasets.transforms": ["numpy", "os"], "modelscope.models.cv.product_retrieval_embedding.item_model": ["numpy", "os", "typing", "torch"], "modelscope.models.cv.product_retrieval_embedding.item_detection": ["numpy", "cv2"], "modelscope.models.cv.product_retrieval_embedding.item_embedding": ["numpy", "cv2", "torch"], "modelscope.models.cv.super_resolution.rrdbnet_arch": ["torch"], "modelscope.models.cv.super_resolution.ecb": ["torch"], "modelscope.models.cv.super_resolution.arch_util": ["collections", "itertools", "torchvision", "math", "warnings", "torch"], "modelscope.models.cv.super_resolution.ecbsr_model": ["os", "typing", "torch"], "modelscope.models.cv.surface_recon_common.fields": ["numpy", "torch"], "modelscope.models.cv.surface_recon_common.dataset": ["numpy", "os", "glob", "cv2", "scipy", "torch"], "modelscope.models.cv.surface_recon_common.utils": ["numpy", "mcubes", "torch"], "modelscope.models.cv.surface_recon_common.surface_recon_common": ["numpy", "os", "trimesh", "torch"], "modelscope.models.cv.surface_recon_common.renderer": ["torch"], "modelscope.models.cv.image_matching.utils.misc": ["yacs"], "modelscope.models.cv.image_matching.quadtree_attention_model": ["numpy", "cv2", "torch", "pathlib", "os"], "modelscope.models.cv.image_matching.loftr_quadtree.loftr": ["einops", "torch"], "modelscope.models.cv.image_matching.loftr_quadtree.backbone.resnet_fpn": ["torch"], "modelscope.models.cv.image_matching.loftr_quadtree.utils.fine_matching": ["math", "kornia", "torch"], "modelscope.models.cv.image_matching.loftr_quadtree.utils.position_encoding": ["math", "torch"], "modelscope.models.cv.image_matching.loftr_quadtree.utils.coarse_matching": ["einops", "torch"], "modelscope.models.cv.image_matching.loftr_quadtree.loftr_module.fine_preprocess": ["einops", "torch"], "modelscope.models.cv.image_matching.loftr_quadtree.loftr_module.transformer": ["math", "timm", "torch", "einops", "copy"], "modelscope.models.cv.image_matching.loftr_quadtree.loftr_module.linear_attention": ["torch"], "modelscope.models.cv.image_matching.loftr_quadtree.loftr_module.quadtree_attention": ["timm", "torch"], "modelscope.models.cv.image_matching.config.default": ["yacs"], "modelscope.models.cv.vop_retrieval.model_se": ["os", "torch"], "modelscope.models.cv.vop_retrieval.backbone": ["numpy", "collections", "tqdm", "warnings", "torch", "typing", "urllib", "hashlib", "os"], "modelscope.models.cv.vop_retrieval.model": ["os", "torch"], "modelscope.models.cv.vop_retrieval.basic_utils": ["numpy", "collections", "torchvision", "ujson", "pickle", "shutil", "cv2", "torch", "PIL", "os", "zipfile", "random"], "modelscope.models.cv.vop_retrieval.tokenization_clip": ["regex", "html", "gzip", "ftfy", "functools", "os", "torch"], "modelscope.models.cv.nerf_recon_4k.nerf_recon_4k": ["numpy", "time", "tqdm", "argparse", "torch", "random", "mmcv", "os", "imageio"], "modelscope.models.cv.nerf_recon_4k.nerf_preprocess": ["numpy", "subprocess", "typing", "cv2", "os", "tensorflow", "glob"], "modelscope.models.cv.nerf_recon_4k.dataloader.load_llff": ["numpy", "os", "imageio", "scipy", "torch"], "modelscope.models.cv.nerf_recon_4k.dataloader.load_data": ["numpy"], "modelscope.models.cv.nerf_recon_4k.dataloader.load_tankstemple": ["numpy", "os", "imageio", "glob"], "modelscope.models.cv.nerf_recon_4k.dataloader.load_blender": ["numpy", "os", "torch", "json", "imageio", "cv2"], "modelscope.models.cv.nerf_recon_4k.dataloader.read_write_model": ["numpy", "collections", "struct", "os", "argparse"], "modelscope.models.cv.nerf_recon_4k.network.utils": ["numpy", "mcubes", "collections", "tinycudann", "gc", "torch"], "modelscope.models.cv.nerf_recon_4k.network.dvgo": ["numpy", "time", "functools", "math", "torch", "torch_scatter", "copy", "os"], "modelscope.models.cv.language_guided_video_summarization.transformer.modules": ["torch"], "modelscope.models.cv.language_guided_video_summarization.transformer.models": ["numpy", "torch"], "modelscope.models.cv.language_guided_video_summarization.transformer.sub_layers": ["numpy", "torch"], "modelscope.models.cv.language_guided_video_summarization.transformer.layers": ["torch"], "modelscope.models.cv.language_guided_video_summarization.summarizer": ["numpy", "videofeatures_clipit", "argparse", "typing", "torch", "bmt_clipit", "os"], "modelscope.models.cv.nerf_recon_vq_compression.utils": ["numpy", "torchvision", "plyfile", "cv2", "skimage", "torch", "PIL", "scipy"], "modelscope.models.cv.nerf_recon_vq_compression.dataloader.blender": ["numpy", "torchvision", "tqdm", "json", "torch", "cv2", "PIL", "os"], "modelscope.models.cv.nerf_recon_vq_compression.dataloader.ray_utils": ["re", "numpy", "kornia", "torch"], "modelscope.models.cv.nerf_recon_vq_compression.dataloader.llff": ["numpy", "torchvision", "torch", "PIL", "os", "glob"], "modelscope.models.cv.nerf_recon_vq_compression.dataloader.nsvf": ["torchvision", "tqdm", "torch", "PIL", "os"], "modelscope.models.cv.nerf_recon_vq_compression.dataloader.tankstemple": ["torchvision", "tqdm", "torch", "PIL", "os"], "modelscope.models.cv.nerf_recon_vq_compression.network.tensoRF_VQ": ["tqdm", "typing", "torch", "os", "random"], "modelscope.models.cv.nerf_recon_vq_compression.network.tensorBase": ["numpy", "time", "torch"], "modelscope.models.cv.nerf_recon_vq_compression.network.tensoRF": [], "modelscope.models.cv.nerf_recon_vq_compression.network.weighted_vq": ["einops", "contextlib", "torch"], "modelscope.models.cv.nerf_recon_vq_compression.renderer": ["numpy", "tqdm", "torch", "sys", "os", "imageio"], "modelscope.models.cv.nerf_recon_vq_compression.nerf_recon_vq_compression": ["numpy", "time", "functools", "tqdm", "cv2", "torch", "os", "glob"], "modelscope.models.cv.image_probing_model.backbone": ["numpy", "collections", "torchvision", "functools", "math", "torch", "operator", "PIL", "sys"], "modelscope.models.cv.image_probing_model.utils": ["re", "torch"], "modelscope.models.cv.image_probing_model.model": ["json", "os", "typing", "torch"], "modelscope.models.cv.table_recognition.modules.lore_detector": ["numpy", "copy", "math", "os", "torch"], "modelscope.models.cv.table_recognition.modules.lore_processor": ["numpy", "copy", "math", "os", "torch"], "modelscope.models.cv.table_recognition.lineless_table_process": ["numpy", "shapely", "torch", "cv2"], "modelscope.models.cv.table_recognition.model_lore": ["numpy", "math", "typing", "torch", "copy", "os"], "modelscope.models.cv.video_panoptic_segmentation.neck.fpn": ["mmcv", "torch"], "modelscope.models.cv.video_panoptic_segmentation.video_k_net": ["mmcv", "numpy", "mmdet", "torch"], "modelscope.models.cv.video_panoptic_segmentation.track.quasi_dense_embed_tracker": ["mmcv", "mmdet", "torch"], "modelscope.models.cv.video_panoptic_segmentation.backbone.swin_checkpoint": ["collections", "torchvision", "importlib", "pkgutil", "os", "torch"], "modelscope.models.cv.video_panoptic_segmentation.backbone.swin_transformer": ["numpy", "mmdet", "timm", "torch"], "modelscope.models.cv.video_panoptic_segmentation.visualizer": ["hashlib", "numpy", "cv2"], "modelscope.models.cv.video_panoptic_segmentation.head.kernel_head": ["mmcv", "torch"], "modelscope.models.cv.video_panoptic_segmentation.head.semantic_fpn_wrapper": ["mmcv", "mmdet", "torch"], "modelscope.models.cv.video_panoptic_segmentation.head.track_heads": ["mmcv", "numpy", "torch"], "modelscope.models.cv.video_panoptic_segmentation.head.kernel_updator": ["mmcv", "torch"], "modelscope.models.cv.video_panoptic_segmentation.head.kernel_update_head": ["mmcv", "numpy", "mmdet", "torch"], "modelscope.models.cv.video_panoptic_segmentation.head.mask": ["numpy", "__future__", "torch", "cv2", "pycocotools"], "modelscope.models.cv.video_panoptic_segmentation.head.kernel_iter_head": ["mmdet", "torch"], "modelscope.models.cv.face_human_hand_detection.shufflenetv2": ["torch"], "modelscope.models.cv.face_human_hand_detection.one_stage_detector": ["torch"], "modelscope.models.cv.face_human_hand_detection.ghost_pan": ["math", "torch"], "modelscope.models.cv.face_human_hand_detection.utils": ["torch"], "modelscope.models.cv.face_human_hand_detection.nanodet_plus_head": ["numpy", "torchvision", "math", "cv2", "torch"], "modelscope.models.cv.face_human_hand_detection.det_infer": ["numpy", "cv2", "torch"], "modelscope.models.cv.image_portrait_enhancement.eqface.model_resnet": ["torch"], "modelscope.models.cv.image_portrait_enhancement.eqface.fqa": ["numpy", "cv2", "os", "torch"], "modelscope.models.cv.image_portrait_enhancement.gpen": ["functools", "math", "torch", "operator", "itertools", "random"], "modelscope.models.cv.image_portrait_enhancement.retinaface.models.net": ["time", "torchvision", "torch"], "modelscope.models.cv.image_portrait_enhancement.retinaface.models.retinaface": ["collections", "torchvision", "torch"], "modelscope.models.cv.image_portrait_enhancement.retinaface.utils": ["numpy", "math", "itertools", "torch"], "modelscope.models.cv.image_portrait_enhancement.retinaface.detection": ["numpy", "cv2", "os", "torch"], "modelscope.models.cv.image_portrait_enhancement.align_faces": ["skimage", "numpy", "cv2"], "modelscope.models.cv.image_portrait_enhancement.losses.model_irse": ["torch"], "modelscope.models.cv.image_portrait_enhancement.losses.losses": ["torch"], "modelscope.models.cv.image_portrait_enhancement.losses.helpers": ["collections", "torch"], "modelscope.models.cv.image_portrait_enhancement.image_portrait_enhancement": ["math", "os", "typing", "torch"], "modelscope.models.cv.image_inpainting.base": ["typing", "torch"], "modelscope.models.cv.image_inpainting.refinement": ["numpy", "tqdm", "kornia", "cv2", "torch"], "modelscope.models.cv.image_inpainting.modules.feature_matching": ["typing", "torch"], "modelscope.models.cv.image_inpainting.modules.ade20k.base": ["os", "torch"], "modelscope.models.cv.image_inpainting.modules.ade20k.resnet": ["math", "os", "torch"], "modelscope.models.cv.image_inpainting.modules.inception": ["torchvision", "torch"], "modelscope.models.cv.image_inpainting.modules.perceptual": ["torchvision", "torch"], "modelscope.models.cv.image_inpainting.modules.ffc": ["numpy", "kornia", "torch"], "modelscope.models.cv.image_inpainting.modules.pix2pixhd": ["numpy", "collections", "functools", "logging", "torch"], "modelscope.models.cv.image_inpainting.modules.adversarial": ["typing", "torch"], "modelscope.models.cv.image_inpainting.model": ["os", "typing", "torch"], "modelscope.models.cv.image_inpainting.default": ["bisect", "torch"], "modelscope.models.cv.open_vocabulary_detection_vild.vild": ["numpy", "typing", "torch", "scipy", "clip", "os", "tensorflow"], "modelscope.models.cv.image_super_resolution_pasd.controlnet": ["torchvision", "dataclasses", "typing", "torch", "copy", "diffusers"], "modelscope.models.cv.image_super_resolution_pasd.transformer_2d": ["dataclasses", "diffusers", "typing", "torch"], "modelscope.models.cv.image_super_resolution_pasd.unet_2d_condition": ["dataclasses", "json", "typing", "torch", "diffusers", "os"], "modelscope.models.cv.image_super_resolution_pasd.attention": ["math", "diffusers", "typing", "torch"], "modelscope.models.cv.image_super_resolution_pasd.misc": ["PIL", "torchvision", "diffusers", "safetensors", "torch"], "modelscope.models.cv.image_super_resolution_pasd.unet_2d_blocks": ["numpy", "diffusers", "typing", "torch"], "modelscope.models.cv.image_quality_assessment_man.maniqa": ["timm", "einops", "torch"], "modelscope.models.cv.image_quality_assessment_man.image_quality_assessment_man": ["os", "typing", "torch"], "modelscope.models.cv.image_quality_assessment_man.swin": ["collections", "itertools", "einops", "math", "warnings", "torch"], "modelscope.models.cv.face_reconstruction.models.facelandmark.nets.large_base_lmks_net": ["torch"], "modelscope.models.cv.face_reconstruction.models.facelandmark.nets.large_eyeball_net": ["torch"], "modelscope.models.cv.face_reconstruction.models.facelandmark.large_base_lmks_infer": ["numpy", "torch"], "modelscope.models.cv.face_reconstruction.models.opt": [], "modelscope.models.cv.face_reconstruction.models.losses": ["numpy", "kornia", "torch"], "modelscope.models.cv.face_reconstruction.models.pix2pix.pix2pix_options": [], "modelscope.models.cv.face_reconstruction.models.pix2pix.pix2pix_model": ["torch"], "modelscope.models.cv.face_reconstruction.models.pix2pix.networks": ["functools", "torch"], "modelscope.models.cv.face_reconstruction.models.nv_diffrast": ["nvdiffrast", "numpy", "warnings", "typing", "torch"], "modelscope.models.cv.face_reconstruction.models.de_retouching_module": ["torch"], "modelscope.models.cv.face_reconstruction.models.facerecon_model": ["numpy", "collections", "cv2", "torch", "os"], "modelscope.models.cv.face_reconstruction.models.unet": ["warnings", "torch"], "modelscope.models.cv.face_reconstruction.models.bfm": ["numpy", "os", "scipy", "torch"], "modelscope.models.cv.face_reconstruction.models.networks": ["os", "kornia", "typing", "torch"], "modelscope.models.cv.face_reconstruction.models.renderer": ["skimage", "numpy", "imageio", "torch"], "modelscope.models.cv.face_reconstruction.utils": ["numpy", "array", "numba", "math", "argparse", "cv2", "torch", "PIL", "os", "scipy"], "modelscope.models.cv.image_reid_person.transreid_model": ["functools", "collections", "itertools", "torch"], "modelscope.models.cv.image_reid_person.pass_model": ["enum", "os", "torch"], "modelscope.models.cv.image_color_enhance.csrnet": ["functools", "math", "torch"], "modelscope.models.cv.image_color_enhance.image_color_enhance": ["os", "typing", "torch"], "modelscope.models.cv.image_color_enhance.deeplpf.deeplpfnet": ["math", "matplotlib", "torch"], "modelscope.models.cv.image_color_enhance.deeplpf.deeplpf_image_color_enhance": ["os", "typing", "torch"], "modelscope.models.cv.image_color_enhance.adaint.adaint": ["torchvision", "numbers", "typing", "torch", "os"], "modelscope.models.cv.salient_detection.models.senet": ["torch"], "modelscope.models.cv.salient_detection.models.modules": ["torch"], "modelscope.models.cv.salient_detection.models.utils": ["torch"], "modelscope.models.cv.salient_detection.models.backbone.Res2Net_v1b": ["math", "torch"], "modelscope.models.cv.salient_detection.models.u2net": ["torch"], "modelscope.models.cv.salient_detection.salient_model": ["torchvision", "cv2", "PIL", "torch", "os"], "modelscope.models.cv.movie_scene_segmentation.utils.save_op": ["numpy", "subprocess", "tqdm", "os", "cv2"], "modelscope.models.cv.movie_scene_segmentation.utils.shot_encoder": ["typing", "torch"], "modelscope.models.cv.movie_scene_segmentation.utils.trn": ["transformers", "torch"], "modelscope.models.cv.movie_scene_segmentation.utils.head": ["torch"], "modelscope.models.cv.movie_scene_segmentation.model": ["numpy", "torchvision", "math", "tqdm", "typing", "torch", "PIL", "einops", "shotdetect_scenedetect_lgss", "os"], "modelscope.models.cv.movie_scene_segmentation.get_model": [], "modelscope.models.cv.video_human_matting.models.matting": ["typing", "torch"], "modelscope.models.cv.video_human_matting.models.deep_guided_filter": ["torch"], "modelscope.models.cv.video_human_matting.models.decoder": ["typing", "torch"], "modelscope.models.cv.video_human_matting.models.lraspp": ["torch"], "modelscope.models.cv.video_human_matting.models.effv2": ["torch"], "modelscope.models.cv.video_human_matting.model": ["numpy", "torchvision", "typing", "torch", "os"], "modelscope.models.cv.text_to_360panorama_image.pipeline_sr": ["numpy", "transformers", "re", "warnings", "inspect", "PIL", "torch", "typing", "copy", "diffusers", "os"], "modelscope.models.cv.text_to_360panorama_image.pipeline_base": ["transformers", "re", "packaging", "typing", "warnings", "inspect", "torch", "diffusers"], "modelscope.models.cv.object_detection.mmdet_ms.dense_heads.rpn_head": ["mmcv", "mmdet", "copy", "torch"], "modelscope.models.cv.object_detection.mmdet_ms.dense_heads.anchor_head": ["mmdet"], "modelscope.models.cv.object_detection.mmdet_ms.roi_heads.bbox_heads.convfc_bbox_head": ["mmdet", "torch"], "modelscope.models.cv.object_detection.mmdet_ms.roi_heads.mask_heads.fcn_mask_head": ["numpy", "mmdet", "warnings", "torch", "mmcv"], "modelscope.models.cv.object_detection.mmdet_ms.utils.convModule_norm": ["mmcv"], "modelscope.models.cv.object_detection.mmdet_ms.utils.checkpoint": ["time", "collections", "torchvision", "warnings", "torch", "tempfile", "importlib", "pkgutil", "mmcv", "os", "io"], "modelscope.models.cv.object_detection.mmdet_ms.backbones.vit": ["functools", "math", "mmdet", "timm", "torch"], "modelscope.models.cv.object_detection.mmdet_ms.necks.fpn": ["mmcv", "mmdet", "torch"], "modelscope.models.cv.object_detection.mmdet_model": ["numpy", "os", "torch"], "modelscope.models.cv.video_object_segmentation.inference_memory_bank": ["math", "torch"], "modelscope.models.cv.video_object_segmentation.modules": ["torchvision", "torch"], "modelscope.models.cv.video_object_segmentation.mod_resnet": ["math", "collections", "torch"], "modelscope.models.cv.video_object_segmentation.cbam": ["torch"], "modelscope.models.cv.video_object_segmentation.aggregate": ["torch"], "modelscope.models.cv.video_object_segmentation.eval_network": ["torch"], "modelscope.models.cv.video_object_segmentation.model": ["os", "typing", "torch"], "modelscope.models.cv.video_object_segmentation.network": ["math", "torch"], "modelscope.models.cv.video_object_segmentation.inference_core": ["torch"], "modelscope.models.cv.image_binary_quant_classification.bnext": ["numpy", "torch"], "modelscope.models.cv.image_binary_quant_classification.binary_quant_model": ["collections", "os", "torch"], "modelscope.models.cv.ocr_detection.preprocessor": ["numpy", "math", "typing", "cv2", "PIL", "torch", "os"], "modelscope.models.cv.ocr_detection.modules.dbnet": ["collections", "math", "torch", "sys", "os"], "modelscope.models.cv.ocr_detection.modules.mix_ops": ["numpy", "math", "torch"], "modelscope.models.cv.ocr_detection.modules.seg_detector_loss": ["sys", "torch"], "modelscope.models.cv.ocr_detection.modules.proxyless": ["numpy", "torch", "sys", "re"], "modelscope.models.cv.ocr_detection.modules.layers": ["numpy", "collections", "torch"], "modelscope.models.cv.ocr_detection.utils": ["pyclipper", "numpy", "shapely", "cv2"], "modelscope.models.cv.ocr_detection.model": ["numpy", "os", "typing", "torch"], "modelscope.models.cv.human_image_generation.human_image_generation_infer": ["numpy", "ast", "torchvision", "math", "pickle", "torch", "cv2", "PIL", "random"], "modelscope.models.cv.human_image_generation.generators.tps": ["torch"], "modelscope.models.cv.human_image_generation.generators.conv2d_gradfix": ["contextlib", "warnings", "torch"], "modelscope.models.cv.human_image_generation.generators.flow_module": ["math", "torch"], "modelscope.models.cv.human_image_generation.generators.wavelet_module": ["numpy", "torch"], "modelscope.models.cv.human_image_generation.generators.base_function": ["collections", "math", "pytorch_wavelets", "torch", "sys"], "modelscope.models.cv.human_image_generation.generators.base_module": ["numpy", "collections", "functools", "math", "tkinter", "torch"], "modelscope.models.cv.human_image_generation.generators.extraction_distribution_model_flow25": ["collections", "sys", "os", "torch"], "modelscope.models.cv.vidt.fpn_fusion": ["torch"], "modelscope.models.cv.vidt.backbone": ["numpy", "os", "math", "timm", "torch"], "modelscope.models.cv.vidt.deformable_transformer": ["copy", "math", "timm", "warnings", "torch"], "modelscope.models.cv.vidt.model": ["os", "torch"], "modelscope.models.cv.vidt.head": ["math", "copy", "torch"], "modelscope.models.cv.image_semantic_segmentation.ddpm_seg.feature_extractors": ["typing", "torch"], "modelscope.models.cv.image_semantic_segmentation.ddpm_seg.utils": ["PIL", "numpy", "random", "torch"], "modelscope.models.cv.image_semantic_segmentation.ddpm_seg.data_util": [], "modelscope.models.cv.image_semantic_segmentation.ddpm_seg.pixel_classifier": ["numpy", "collections", "torch", "PIL", "os"], "modelscope.models.cv.image_semantic_segmentation.pan_merge.maskformer_semantic_head": ["mmdet", "torch"], "modelscope.models.cv.image_semantic_segmentation.pan_merge.base_panoptic_fusion_head": ["mmcv", "mmdet", "abc"], "modelscope.models.cv.image_semantic_segmentation.semantic_seg_model": ["numpy", "os", "torch"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.segmentors.encoder_decoder_mask2former": ["mmdet", "torch"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.segmentors.base_segmentor": ["numpy", "collections", "mmcv", "abc", "warnings", "torch"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.decode_heads.base_decode_head": ["mmcv", "abc", "mmdet", "torch"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.decode_heads.mask2former_head_from_mmseg": ["mmcv", "mmdet", "copy", "torch"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.backbone.beit_adapter": ["mmdet", "math", "timm", "torch", "logging"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.backbone.adapter_modules": ["logging", "functools", "mmdet", "timm", "torch"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.models.backbone.base.beit": ["mmcv", "functools", "math", "mmdet", "timm", "torch"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.utils.seg_func": ["warnings", "torch"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.utils.builder": ["mmcv"], "modelscope.models.cv.image_semantic_segmentation.vit_adapter.utils.data_process_func": ["mmcv", "mmdet"], "modelscope.models.cv.image_semantic_segmentation.ddpm_segmentation_model": ["ddpm_guided_diffusion", "os", "typing", "torch"], "modelscope.models.cv.image_classification.utils": ["numpy", "collections", "itertools", "mmcls", "math", "os", "torch"], "modelscope.models.cv.image_classification.resnet50_cc": ["collections", "torchvision", "math", "torch", "os"], "modelscope.models.cv.image_classification.backbones.beit_v2": ["collections", "functools", "math", "warnings", "torch", "typing", "itertools", "einops", "mmcls", "mmcv", "os"], "modelscope.models.cv.image_classification.backbones.nextvit": ["collections", "functools", "math", "warnings", "torch", "typing", "itertools", "einops", "mmcls", "mmcv", "os"], "modelscope.models.cv.image_classification.mmcls_model": ["os"], "modelscope.models.cv.image_denoise.nafnet_for_image_denoise": ["os", "typing", "torch"], "modelscope.models.cv.image_denoise.nafnet.NAFNet_arch": ["numpy", "torch"], "modelscope.models.cv.image_denoise.nafnet.arch_util": ["torch"], "modelscope.models.cv.vision_efficient_tuning.timm_vision_transformer": ["collections", "functools", "math", "torch", "itertools", "logging"], "modelscope.models.cv.vision_efficient_tuning.timm_weight_init": ["math", "warnings", "torch"], "modelscope.models.cv.vision_efficient_tuning.backbone": ["functools", "torch"], "modelscope.models.cv.vision_efficient_tuning.petl": ["math", "collections", "torchvision", "torch"], "modelscope.models.cv.vision_efficient_tuning.model": ["typing", "torch"], "modelscope.models.cv.vision_efficient_tuning.vision_efficient_tuning": ["collections", "os", "torch"], "modelscope.models.cv.vision_efficient_tuning.timm_helpers": ["math", "itertools", "typing", "torch"], "modelscope.models.cv.vision_efficient_tuning.head": ["torch"], "modelscope.models.cv.body_2d_keypoints.hrnet_v2": ["numpy", "os", "torch"], "modelscope.models.cv.body_2d_keypoints.hrnet_basic_modules": ["torch"], "modelscope.models.cv.body_2d_keypoints.w48": [], "modelscope.models.cv.video_single_object_tracking.models.layers.attn_blocks": ["math", "timm", "torch"], "modelscope.models.cv.video_single_object_tracking.models.layers.attn": ["torch"], "modelscope.models.cv.video_single_object_tracking.models.layers.patch_embed": ["timm", "torch"], "modelscope.models.cv.video_single_object_tracking.models.layers.head": ["torch"], "modelscope.models.cv.video_single_object_tracking.models.procontext.vit_ce": ["functools", "timm", "torch"], "modelscope.models.cv.video_single_object_tracking.models.procontext.utils": ["torch"], "modelscope.models.cv.video_single_object_tracking.models.procontext.procontext": ["torch"], "modelscope.models.cv.video_single_object_tracking.models.ostrack.vit_ce": ["functools", "timm", "torch"], "modelscope.models.cv.video_single_object_tracking.models.ostrack.utils": ["torch"], "modelscope.models.cv.video_single_object_tracking.models.ostrack.base_backbone": ["timm", "torch"], "modelscope.models.cv.video_single_object_tracking.models.ostrack.ostrack": ["torch"], "modelscope.models.cv.video_single_object_tracking.utils.utils": ["numpy", "math", "cv2", "typing", "torch"], "modelscope.models.cv.video_single_object_tracking.config.ostrack": ["easydict"], "modelscope.models.cv.video_single_object_tracking.tracker.ostrack": ["torch"], "modelscope.models.cv.video_single_object_tracking.tracker.procontext": ["torch", "copy"], "modelscope.models.cv.cartoon.loss": ["numpy", "skimage", "tensorflow", "joblib", "os", "scipy"], "modelscope.models.cv.cartoon.utils": ["numpy", "random", "os", "tensorflow", "cv2"], "modelscope.models.cv.cartoon.model_tf": ["tensorflow", "typing"], "modelscope.models.cv.cartoon.facelib.facer": ["numpy", "time", "cv2"], "modelscope.models.cv.cartoon.facelib.face_landmark": ["numpy", "tensorflow", "cv2"], "modelscope.models.cv.cartoon.facelib.config": ["numpy", "easydict", "os"], "modelscope.models.cv.cartoon.facelib.face_detector": ["numpy", "time", "tensorflow", "cv2"], "modelscope.models.cv.cartoon.facelib.LK.lk": ["numpy"], "modelscope.models.cv.cartoon.network": ["tensorflow"], "modelscope.models.cv.cartoon.mtcnn_pytorch.src.matlab_cp2tform": ["numpy"], "modelscope.models.cv.cartoon.mtcnn_pytorch.src.align_trans": ["numpy", "cv2"], "modelscope.models.cv.controllable_image_generation.annotator.openpose.body": ["numpy", "time", "torchvision", "math", "cv2", "torch", "matplotlib", "scipy"], "modelscope.models.cv.controllable_image_generation.annotator.openpose.hand": ["numpy", "time", "math", "json", "cv2", "torch", "skimage", "matplotlib", "scipy"], "modelscope.models.cv.controllable_image_generation.annotator.openpose.model": ["collections", "torch"], "modelscope.models.cv.controllable_image_generation.annotator.openpose.util": ["numpy", "math", "matplotlib", "cv2"], "modelscope.models.cv.controllable_image_generation.annotator.annotator": ["numpy", "mmseg", "torch", "cv2", "einops", "mmcv", "os"], "modelscope.models.cv.controllable_image_generation.annotator.mlsd.mbv2_mlsd_large": ["os", "sys", "torch"], "modelscope.models.cv.controllable_image_generation.annotator.mlsd.utils": ["numpy", "torch", "os", "cv2"], "modelscope.models.cv.controllable_image_generation.annotator.midas.api": ["cv2", "torchvision", "os", "torch"], "modelscope.models.cv.controllable_image_generation.annotator.midas.utils": ["numpy", "sys", "re", "cv2", "torch"], "modelscope.models.cv.controllable_image_generation.annotator.midas.midas.transforms": ["numpy", "math", "cv2"], "modelscope.models.cv.controllable_image_generation.annotator.midas.midas.blocks": ["torch"], "modelscope.models.cv.controllable_image_generation.annotator.midas.midas.base_model": ["torch"], "modelscope.models.cv.controllable_image_generation.annotator.midas.midas.vit": ["timm", "math", "types", "torch"], "modelscope.models.cv.controllable_image_generation.annotator.midas.midas.midas_net_custom": ["torch"], "modelscope.models.cv.controllable_image_generation.annotator.midas.midas.dpt_depth": ["torch"], "modelscope.models.cv.controllable_image_generation.annotator.midas.midas.midas_net": ["torch"], "modelscope.models.cv.controllable_image_generation.controlnet": ["numpy", "control_ldm", "math", "typing", "cv2", "torch", "PIL", "tempfile", "sys", "einops", "os", "random"], "modelscope.models.cv.video_inpainting.inpainting_model": ["numpy", "math", "torchvision", "torch"], "modelscope.models.cv.video_inpainting.inpainting": ["PIL", "numpy", "time", "torchvision", "os", "cv2", "torch"], "modelscope.models.cv.image_body_reshaping.person_info": ["numpy", "copy", "cv2", "torch"], "modelscope.models.cv.image_body_reshaping.image_body_reshaping": ["numpy", "typing", "cv2", "torch", "os"], "modelscope.models.cv.image_body_reshaping.slim_utils": ["numpy", "numba", "random", "math", "torch", "os", "cv2"], "modelscope.models.cv.image_body_reshaping.model": ["torch"], "modelscope.models.cv.image_body_reshaping.pose_estimator.body": ["numpy", "math", "cv2", "torch", "scipy"], "modelscope.models.cv.image_body_reshaping.pose_estimator.model": ["collections", "torch"], "modelscope.models.cv.image_body_reshaping.pose_estimator.util": ["numpy"], "modelscope.models.cv.video_instance_segmentation.neck.msdeformattn_decoder": ["mmcv", "mmdet", "torch"], "modelscope.models.cv.video_instance_segmentation.track.kernel_update_head": ["mmcv", "numpy", "mmdet", "torch"], "modelscope.models.cv.video_instance_segmentation.track.mask_hungarian_assigner": ["numpy", "mmdet", "scipy", "torch"], "modelscope.models.cv.video_instance_segmentation.video_knet": ["mmdet", "torch"], "modelscope.models.cv.video_instance_segmentation.utils": ["numpy", "mmdet", "torch"], "modelscope.models.cv.video_instance_segmentation.head.kernel_head": ["mmcv", "mmdet", "torch"], "modelscope.models.cv.video_instance_segmentation.head.kernel_frame_iter_head": ["mmcv", "mmdet", "torch"], "modelscope.models.cv.video_instance_segmentation.head.kernel_updator": ["mmcv", "torch"], "modelscope.models.cv.video_instance_segmentation.head.kernel_update_head": ["mmcv", "numpy", "mmdet", "torch"], "modelscope.models.cv.video_instance_segmentation.head.kernel_iter_head": ["mmdet", "torch"], "modelscope.models.cv.image_editing.masactrl": ["einops", "torch"], "modelscope.models.cv.image_editing.masactrl_utils": ["einops", "torch"], "modelscope.models.cv.vision_middleware.vim": ["math", "einops", "torch"], "modelscope.models.cv.vision_middleware.backbone": ["numpy", "collections", "math", "typing", "torch", "os"], "modelscope.models.cv.vision_middleware.model": ["json", "os", "typing", "torch"], "modelscope.models.cv.vision_middleware.head": ["mmcv", "numpy", "abc", "torch"], "modelscope.models.cv.video_multi_object_tracking.models.decode": ["torch"], "modelscope.models.cv.video_multi_object_tracking.models.model": ["torch"], "modelscope.models.cv.video_multi_object_tracking.models.common": ["torch"], "modelscope.models.cv.video_multi_object_tracking.models.yolo": ["math", "copy", "torch"], "modelscope.models.cv.video_multi_object_tracking.utils.kalman_filter": ["numpy", "scipy"], "modelscope.models.cv.video_multi_object_tracking.utils.visualization": ["numpy", "cv2"], "modelscope.models.cv.video_multi_object_tracking.utils.image": ["numpy", "cv2"], "modelscope.models.cv.video_multi_object_tracking.utils.utils": ["numpy", "cv2", "torch"], "modelscope.models.cv.video_multi_object_tracking.tracker.matching": ["numpy", "lap", "scipy"], "modelscope.models.cv.video_multi_object_tracking.tracker.multitracker": ["numpy", "collections", "torch"], "modelscope.models.cv.video_multi_object_tracking.tracker.basetrack": ["numpy", "collections"], "modelscope.models.cv.action_recognition.models": ["torch"], "modelscope.models.cv.action_recognition.tada_convnext": ["math", "torch"], "modelscope.models.cv.action_recognition.s3dg": ["torch"], "modelscope.models.cv.action_recognition.temporal_patch_shift_transformer": ["numpy", "torchvision", "functools", "abc", "timm", "torch", "operator", "einops"], "modelscope.models.cv.image_defrcn_fewshot.evaluation.pascal_voc_evaluation": ["numpy", "collections", "tempfile", "detectron2", "os"], "modelscope.models.cv.image_defrcn_fewshot.evaluation.coco_evaluation": ["numpy", "collections", "json", "contextlib", "torch", "pycocotools", "tabulate", "itertools", "fvcore", "copy", "detectron2", "logging", "os", "io"], "modelscope.models.cv.image_defrcn_fewshot.evaluation.evaluator": ["time", "torch", "datetime", "detectron2", "logging"], "modelscope.models.cv.image_defrcn_fewshot.models.defrcn": ["detectron2", "os", "typing", "torch"], "modelscope.models.cv.image_defrcn_fewshot.models.fast_rcnn": ["detectron2", "numpy", "fvcore", "torch"], "modelscope.models.cv.image_defrcn_fewshot.models.roi_heads": ["detectron2", "torch"], "modelscope.models.cv.image_defrcn_fewshot.models.gdl": ["torch"], "modelscope.models.cv.image_defrcn_fewshot.models.resnet": ["torchvision", "torch"], "modelscope.models.cv.image_defrcn_fewshot.models.calibration_layer": ["detectron2", "cv2", "sklearn", "torch"], "modelscope.models.cv.image_defrcn_fewshot.utils.requirements_check": ["collections", "packaging", "importlib_metadata", "sys", "importlib"], "modelscope.models.cv.image_defrcn_fewshot.utils.voc_register": ["numpy", "detectron2", "fvcore", "xml", "os"], "modelscope.models.cv.image_defrcn_fewshot.utils.coco_register": ["fvcore", "detectron2", "io", "os", "contextlib", "pycocotools"], "modelscope.models.cv.image_defrcn_fewshot.utils.configuration_mapper": ["detectron2"], "modelscope.models.cv.image_defrcn_fewshot.utils.model_surgery_op": ["os", "argparse", "torch"], "modelscope.models.cv.image_defrcn_fewshot.utils.register_data": [], "modelscope.models.cv.image_defrcn_fewshot.defrcn_for_fewshot": ["os", "typing", "torch"], "modelscope.models.cv.image_to_image_translation.model_translation": ["math", "torch"], "modelscope.models.cv.image_to_image_translation.models.autoencoder": ["math", "torch"], "modelscope.models.cv.image_to_image_translation.models.clip": ["math", "torch"], "modelscope.models.cv.image_to_image_translation.ops.apps": ["numpy", "torchvision", "torch", "PIL", "os", "artist"], "modelscope.models.cv.image_to_image_translation.ops.diffusion": ["math", "torch"], "modelscope.models.cv.image_to_image_translation.ops.losses": ["math", "torch"], "modelscope.models.cv.image_to_image_translation.ops.utils": ["numpy", "math", "json", "binascii", "cv2", "base64", "torch", "PIL", "zipfile", "hashlib", "multiprocessing", "os", "io"], "modelscope.models.cv.image_to_image_translation.ops.degradation": ["numpy", "random", "math", "torch", "os", "scipy", "cv2"], "modelscope.models.cv.image_to_image_translation.ops.svd": ["torch"], "modelscope.models.cv.image_to_image_translation.ops.random_mask": ["numpy", "cv2"], "modelscope.models.cv.image_to_image_translation.ops.random_color": ["colorsys", "random"], "modelscope.models.cv.image_to_image_translation.ops.metrics": ["numpy", "scipy", "torch"], "modelscope.models.cv.image_to_image_translation.data.transforms": ["PIL", "math", "random", "torchvision"], "modelscope.models.cv.stream_yolo.exp.base_exp": ["abc", "torch"], "modelscope.models.cv.stream_yolo.exp.build": ["os", "sys"], "modelscope.models.cv.stream_yolo.exp.default.streamyolo": ["os", "sys", "torch"], "modelscope.models.cv.stream_yolo.exp.yolox_base": ["random", "os", "torch"], "modelscope.models.cv.stream_yolo.models.streamyolo": ["torch"], "modelscope.models.cv.stream_yolo.models.darknet": ["torch"], "modelscope.models.cv.stream_yolo.models.tal_head": ["torch"], "modelscope.models.cv.stream_yolo.models.dfp_pafpn": ["torch"], "modelscope.models.cv.stream_yolo.models.network_blocks": ["torch"], "modelscope.models.cv.stream_yolo.realtime_video_detector": ["numpy", "time", "tqdm", "json", "argparse", "cv2", "torch", "logging", "os"], "modelscope.models.cv.stream_yolo.utils.boxes": ["torchvision", "torch"], "modelscope.models.cv.stream_yolo.utils.format": ["math"], "modelscope.models.cv.stream_yolo.data.data_augment": ["numpy", "math", "random", "cv2"], "modelscope.models.cv.abnormal_object_detection.mmdet_ms.roi_head.roi_extractors.single_level_roi_extractor": ["mmcv", "mmdet", "torch"], "modelscope.models.cv.abnormal_object_detection.mmdet_ms.roi_head.mask_scoring_roi_head": ["mmdet", "torch"], "modelscope.models.cv.abnormal_object_detection.mmdet_model": ["numpy", "os", "torch"], "modelscope.models.cv.action_detection.modules.action_detection_pytorch": ["typing", "torch", "fvcore", "detectron2", "logging"], "modelscope.models.cv.action_detection.modules.resnet": ["detectron2", "torch"], "modelscope.models.cv.action_detection.action_detection_onnx": ["numpy", "onnxruntime", "subprocess", "shutil", "cv2", "urllib", "uuid", "tempfile", "os"], "modelscope.models.cv.image_restoration.image_restoration_model": ["numpy", "cv2", "os", "torch"], "modelscope.models.cv.image_restoration.demoire_models.nets": ["torch"], "modelscope.models.cv.hand_static.hand_model": ["numpy", "torchvision", "cv2", "PIL", "torch", "sys", "os"], "modelscope.models.cv.hand_static.networks": ["os", "torchvision", "torch"], "modelscope.models.cv.indoor_layout_estimation.panovit": ["yacs", "numpy", "os", "torch"], "modelscope.models.cv.indoor_layout_estimation.networks.panovit": ["numpy", "torch"], "modelscope.models.cv.indoor_layout_estimation.networks.modality.layout": ["numpy", "math", "torch", "scipy", "shapely"], "modelscope.models.cv.indoor_layout_estimation.networks.misc.fourier": ["PIL", "numpy", "scipy"], "modelscope.models.cv.indoor_layout_estimation.networks.misc.post_proc": ["numpy", "sklearn", "scipy"], "modelscope.models.cv.indoor_layout_estimation.networks.misc.panostretch": ["functools", "numpy", "scipy"], "modelscope.models.cv.indoor_layout_estimation.networks.utils": ["numpy", "torch"], "modelscope.models.cv.indoor_layout_estimation.networks.backbone.vit_horizon_pry_image": ["numpy", "timm", "torch"], "modelscope.models.cv.indoor_layout_estimation.networks.backbone.resnet_DA": ["torchvision", "torch"], "modelscope.models.cv.nerf_recon_acc.nerf_preprocess": ["numpy", "subprocess", "typing", "cv2", "os", "tensorflow", "glob"], "modelscope.models.cv.nerf_recon_acc.dataloader.nerf_dataset": ["numpy", "torchvision", "math", "json", "torch", "PIL", "os"], "modelscope.models.cv.nerf_recon_acc.dataloader.read_write_model": ["numpy", "collections", "struct", "os", "argparse"], "modelscope.models.cv.nerf_recon_acc.network.utils": ["numpy", "mcubes", "collections", "tinycudann", "gc", "torch"], "modelscope.models.cv.nerf_recon_acc.network.segmenter": ["numpy", "tensorflow"], "modelscope.models.cv.nerf_recon_acc.network.nerf": ["tinycudann", "numpy", "nerfacc", "torch"], "modelscope.models.cv.nerf_recon_acc.nerf_recon_acc": ["numpy", "time", "tqdm", "cv2", "torch", "os", "glob"], "modelscope.models.cv.image_mvs_depth_estimation.casmvs_model": ["numpy", "cv2", "torch", "easydict", "os"], "modelscope.models.cv.image_mvs_depth_estimation.depth_filter": ["numpy", "plyfile", "cv2", "PIL", "os"], "modelscope.models.cv.image_mvs_depth_estimation.utils": ["numpy", "random", "torchvision", "torch"], "modelscope.models.cv.image_mvs_depth_estimation.cas_mvsnet": ["torch"], "modelscope.models.cv.image_mvs_depth_estimation.colmap2mvsnet": ["numpy", "__future__", "collections", "functools", "shutil", "cv2", "struct", "multiprocessing", "os"], "modelscope.models.cv.image_mvs_depth_estimation.module": ["torch"], "modelscope.models.cv.image_mvs_depth_estimation.general_eval_dataset": ["PIL", "numpy", "sys", "re", "torch", "os", "cv2"], "modelscope.models.cv.referring_video_object_segmentation.utils.position_encoding_2d": ["math", "torch"], "modelscope.models.cv.referring_video_object_segmentation.utils.segmentation": ["typing", "torch"], "modelscope.models.cv.referring_video_object_segmentation.utils.criterion": ["torch"], "modelscope.models.cv.referring_video_object_segmentation.utils.backbone": ["einops", "torchvision", "torch"], "modelscope.models.cv.referring_video_object_segmentation.utils.swin_transformer": ["numpy", "einops", "functools", "torch", "timm", "operator"], "modelscope.models.cv.referring_video_object_segmentation.utils.postprocessing": ["numpy", "torch", "einops", "pycocotools"], "modelscope.models.cv.referring_video_object_segmentation.utils.multimodal_transformer": ["transformers", "typing", "torch", "einops", "copy", "os"], "modelscope.models.cv.referring_video_object_segmentation.utils.mttr": ["einops", "torch"], "modelscope.models.cv.referring_video_object_segmentation.utils.misc": ["pickle", "torchvision", "typing", "torch"], "modelscope.models.cv.referring_video_object_segmentation.utils.matcher": ["scipy", "torch"], "modelscope.models.cv.referring_video_object_segmentation.model": ["os", "typing", "torch"], "modelscope.models.cv.human_reconstruction.models.human_segmenter": ["numpy", "tensorflow", "cv2"], "modelscope.models.cv.human_reconstruction.models.PixToMesh": ["torch"], "modelscope.models.cv.human_reconstruction.models.detectors": ["numpy", "torch"], "modelscope.models.cv.human_reconstruction.models.Embedding": ["torch"], "modelscope.models.cv.human_reconstruction.models.networks": ["functools", "numpy", "torch"], "modelscope.models.cv.human_reconstruction.models.Res_backbone": ["numpy", "torch"], "modelscope.models.cv.human_reconstruction.models.geometry": ["torch"], "modelscope.models.cv.human_reconstruction.models.Surface_head": ["torch"], "modelscope.models.cv.human_reconstruction.utils": ["numpy", "mcubes", "os", "torch"], "modelscope.models.cv.human_reconstruction.Reconstruction": ["numpy", "torchvision", "typing", "torch", "PIL", "cv2", "skimage", "os"], "modelscope.models.cv.image_depth_estimation_bts.depth_estimation_bts_model": ["os", "torch"], "modelscope.models.cv.image_depth_estimation_bts.networks.encoder": ["torchvision", "torch"], "modelscope.models.cv.image_depth_estimation_bts.networks.bts_model": ["torch"], "modelscope.models.cv.image_depth_estimation_bts.networks.utils": ["math", "torch"], "modelscope.models.cv.image_depth_estimation_bts.networks.decoder": ["torch"], "modelscope.models.cv.image_driving_perception.preprocessor": ["numpy", "cv2", "typing", "torch"], "modelscope.models.cv.image_driving_perception.utils": ["numpy", "time", "torchvision", "torch"], "modelscope.models.cv.image_driving_perception.image_driving_percetion_model": ["numpy", "typing", "cv2", "torch", "os"], "modelscope.models.cv.facial_expression_recognition.fer.transforms": ["PIL", "numpy", "types", "numbers", "torch"], "modelscope.models.cv.facial_expression_recognition.fer.vgg": ["torch"], "modelscope.models.cv.facial_expression_recognition.fer.facial_expression_recognition": ["numpy", "cv2", "PIL", "torch", "os"], "modelscope.models.cv.image_human_parsing.m2fp.m2fp_decoder": ["torch"], "modelscope.models.cv.image_human_parsing.m2fp.m2fp_encoder": ["numpy", "typing", "torch"], "modelscope.models.cv.image_human_parsing.backbone.deeplab_resnet": ["numpy", "torch"], "modelscope.models.cv.image_human_parsing.m2fp_net": ["os", "typing", "torch"], "modelscope.models.cv.image_human_parsing.parsing_utils": ["PIL", "numpy", "torch", "copy"], "modelscope.models.cv.shop_segmentation.models": ["math", "collections", "timm", "torch"], "modelscope.models.cv.shop_segmentation.shop_seg_model": ["numpy", "json", "typing", "torch", "PIL", "os"], "modelscope.models.cv.shop_segmentation.utils": ["regex", "html", "gzip", "functools", "typing", "torch", "ftfy", "os"], "modelscope.models.cv.shop_segmentation.head_fpn": ["mmcv", "numpy", "timm", "torch"], "modelscope.models.cv.shop_segmentation.shop_seg_base": ["torch"], "modelscope.models.cv.shop_segmentation.common": ["warnings", "torch"], "modelscope.models.cv.shop_segmentation.neck_fpn": ["mmcv", "timm", "torch"], "modelscope.models.cv.object_detection_3d.depe.result_vis": ["numpy", "pickle", "json", "argparse", "mmdet3d", "cv2", "pyquaternion", "os"], "modelscope.models.cv.object_detection_3d.depe.depe_detect": ["numpy", "os", "typing", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.core.bbox.assigners.hungarian_assigner_3d": ["mmdet", "scipy", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.core.bbox.util": ["numpy", "mmdet3d", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.core.bbox.match_costs.match_cost": ["mmdet", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.core.bbox.coders.nms_free_coder": ["mmdet", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.detectors.petr3d": ["numpy", "mmdet", "mmdet3d", "torch", "mmcv"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.dense_heads.petrv2_dednhead": ["numpy", "mmdet", "math", "mmdet3d", "torch", "copy", "mmcv"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.dense_heads.depth_net": ["mmcv", "mmdet", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.petr_transformer": ["copy", "mmcv", "mmdet", "math", "warnings", "typing", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.utils.positional_encoding": ["mmcv", "math", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.backbones.vovnet": ["mmcv", "collections", "mmdet", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.models.necks.cp_fpn": ["mmcv", "mmdet", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.pipelines.transform_3d": ["PIL", "numpy", "copy", "mmcv", "mmdet", "mmdet3d", "torch"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.pipelines.loading": ["mmcv", "mmdet", "numpy"], "modelscope.models.cv.object_detection_3d.depe.mmdet3d_plugin.datasets.nuscenes_dataset": ["numpy", "mmdet", "mmdet3d"], "modelscope.models.cv.video_depth_estimation.models.model_utils": [], "modelscope.models.cv.video_depth_estimation.models.model_wrapper": ["numpy", "collections", "torch", "importlib", "random"], "modelscope.models.cv.video_depth_estimation.models.sfm_model_mf": ["random", "torch"], "modelscope.models.cv.video_depth_estimation.models.sup_model_mf": [], "modelscope.models.cv.video_depth_estimation.models.model_checkpoint": ["re", "numpy", "os", "torch"], "modelscope.models.cv.video_depth_estimation.networks.layers.resnet.pose_decoder": ["__future__", "collections", "torch"], "modelscope.models.cv.video_depth_estimation.networks.layers.resnet.resnet_encoder": ["numpy", "__future__", "torchvision", "torch"], "modelscope.models.cv.video_depth_estimation.networks.layers.resnet.depth_decoder": ["numpy", "__future__", "collections", "torch"], "modelscope.models.cv.video_depth_estimation.networks.layers.resnet.layers": ["__future__", "torch"], "modelscope.models.cv.video_depth_estimation.networks.depth_pose.depth_pose_net": ["functools", "torch"], "modelscope.models.cv.video_depth_estimation.networks.optim.extractor": ["torchvision", "torch"], "modelscope.models.cv.video_depth_estimation.networks.optim.update": ["torch"], "modelscope.models.cv.video_depth_estimation.utils.augmentations": ["numpy", "torchvision", "cv2", "PIL", "random"], "modelscope.models.cv.video_depth_estimation.utils.load": ["collections", "warnings", "torch", "inspect", "importlib", "logging", "os"], "modelscope.models.cv.video_depth_estimation.utils.image_gt": ["PIL", "functools", "cv2", "torch"], "modelscope.models.cv.video_depth_estimation.utils.types": ["yacs", "numpy", "torch"], "modelscope.models.cv.video_depth_estimation.utils.horovod": ["horovod"], "modelscope.models.cv.video_depth_estimation.utils.image": ["numpy", "functools", "cv2", "PIL", "torch", "os"], "modelscope.models.cv.video_depth_estimation.utils.config": ["yacs", "os", "datetime", "torch"], "modelscope.models.cv.video_depth_estimation.utils.misc": ["termcolor"], "modelscope.models.cv.video_depth_estimation.utils.depth": ["numpy", "matplotlib", "torchvision", "torch"], "modelscope.models.cv.video_depth_estimation.dro_model": ["numpy", "tqdm", "cv2", "torch", "os", "glob"], "modelscope.models.cv.video_depth_estimation.configs.default_config": ["yacs", "os"], "modelscope.models.cv.video_depth_estimation.geometry.pose": ["torch"], "modelscope.models.cv.video_depth_estimation.geometry.camera": ["functools", "torch"], "modelscope.models.cv.video_depth_estimation.geometry.camera_utils": ["torch"], "modelscope.models.cv.video_depth_estimation.geometry.pose_utils": ["numpy", "torch"], "modelscope.models.cv.robust_image_classification.easyrobust_model": ["os", "torch"], "modelscope.models.cv.video_streaming_perception.longshortnet.exp.longshortnet_base": [], "modelscope.models.cv.video_streaming_perception.longshortnet.models.dfp_pafpn_long": ["collections", "torch"], "modelscope.models.cv.video_streaming_perception.longshortnet.models.longshort": ["torch"], "modelscope.models.cv.video_streaming_perception.longshortnet.models.longshort_backbone_neck": ["torch"], "modelscope.models.cv.video_streaming_perception.longshortnet.models.dfp_pafpn_short": ["collections", "torch"], "modelscope.models.cv.video_streaming_perception.longshortnet.longshortnet": ["numpy", "time", "tqdm", "json", "argparse", "cv2", "torch", "logging", "os"], "modelscope.models.cv.video_deinterlace.models.enh": ["torch"], "modelscope.models.cv.video_deinterlace.models.archs": ["numpy", "torch"], "modelscope.models.cv.video_deinterlace.models.utils": ["torch"], "modelscope.models.cv.video_deinterlace.models.fre": ["torch"], "modelscope.models.cv.video_deinterlace.models.deep_fourier_upsampling": ["numpy", "torch"], "modelscope.models.cv.video_deinterlace.UNet_for_video_deinterlace": ["copy", "os", "typing", "torch"], "modelscope.models.cv.video_deinterlace.deinterlace_arch": ["torch"], "modelscope.models.cv.s2net_panorama_depth_estimation.s2net_model": ["numpy", "os", "torchvision", "torch"], "modelscope.models.cv.s2net_panorama_depth_estimation.networks.config": ["yaml", "yacs", "os"], "modelscope.models.cv.s2net_panorama_depth_estimation.networks.swin_transformer": ["numpy", "timm", "torch"], "modelscope.models.cv.s2net_panorama_depth_estimation.networks.model": ["numpy", "torch"], "modelscope.models.cv.s2net_panorama_depth_estimation.networks.resnet": ["torch"], "modelscope.models.cv.s2net_panorama_depth_estimation.networks.decoder": ["numpy", "einops", "torch"], "modelscope.models.cv.s2net_panorama_depth_estimation.networks.util_helper": ["numpy", "apex", "torchvision", "healpy", "warnings", "torch", "importlib", "pkgutil", "os"], "modelscope.models.cv.text_driven_segmentation.lseg_blocks": ["torch"], "modelscope.models.cv.text_driven_segmentation.lseg_base": ["torch"], "modelscope.models.cv.text_driven_segmentation.lseg_model": ["numpy", "json", "typing", "torch", "PIL", "os"], "modelscope.models.cv.text_driven_segmentation.lseg_net": ["numpy", "torch"], "modelscope.models.cv.text_driven_segmentation.model": ["numpy", "collections", "typing", "torch"], "modelscope.models.cv.text_driven_segmentation.lseg_vit": ["types", "math", "timm", "torch"], "modelscope.models.cv.text_driven_segmentation.clip": ["torchvision", "tqdm", "warnings", "torch", "typing", "urllib", "PIL", "pkg_resources", "hashlib", "os"], "modelscope.models.cv.text_driven_segmentation.simple_tokenizer": ["regex", "html", "gzip", "ftfy", "functools", "os"], "modelscope.models.cv.image_face_fusion.image_face_fusion": ["numpy", "collections", "torchvision", "typing", "torch", "PIL", "cv2", "os"], "modelscope.models.cv.image_face_fusion.facelib.matlab_cp2tform": ["numpy"], "modelscope.models.cv.image_face_fusion.facelib.align_trans": ["numpy", "cv2"], "modelscope.models.cv.image_face_fusion.facegan.op.conv2d_gradfix": ["contextlib", "warnings", "torch"], "modelscope.models.cv.image_face_fusion.facegan.op.fused_act": ["torch"], "modelscope.models.cv.image_face_fusion.facegan.op.upfirdn2d": ["collections", "torch"], "modelscope.models.cv.image_face_fusion.facegan.face_gan": ["numpy", "os", "torch"], "modelscope.models.cv.image_face_fusion.facegan.gpen_model": ["functools", "math", "torch", "operator", "itertools", "random"], "modelscope.models.cv.image_face_fusion.network.aei_flow_net": ["torch"], "modelscope.models.cv.image_face_fusion.network.model_irse": ["collections", "torch"], "modelscope.models.cv.image_face_fusion.network.aad_layer": ["torch"], "modelscope.models.cv.image_face_fusion.network.facerecon_model": ["os", "typing", "torch"], "modelscope.models.cv.image_face_fusion.network.bfm": ["numpy", "os", "scipy", "torch"], "modelscope.models.cv.image_face_fusion.network.dense_motion": ["torch"], "modelscope.models.cv.image_face_fusion.network.ops": ["torch"], "modelscope.models.cv.image_deblur.nafnet_for_image_deblur": ["os", "typing", "torch"], "modelscope.models.cv.body_3d_keypoints.cannonical_pose.body_3d_pose": ["numpy", "typing", "torch", "logging", "os"], "modelscope.models.cv.body_3d_keypoints.cannonical_pose.canonical_pose_modules": ["torch"], "modelscope.models.cv.body_3d_keypoints.hdformer.backbone": ["torch"], "modelscope.models.cv.body_3d_keypoints.hdformer.hdformer": ["torch"], "modelscope.models.cv.body_3d_keypoints.hdformer.block": ["math", "einops", "torch"], "modelscope.models.cv.body_3d_keypoints.hdformer.skeleton": ["numpy"], "modelscope.models.cv.body_3d_keypoints.hdformer.directed_graph": ["numpy", "sys", "typing"], "modelscope.models.cv.body_3d_keypoints.hdformer.hdformer_detector": ["numpy", "os", "typing", "torch"], "modelscope.models.cv.video_stabilization.DUT.RAFT.extractor": ["torch"], "modelscope.models.cv.video_stabilization.DUT.RAFT.update": ["torch"], "modelscope.models.cv.video_stabilization.DUT.RAFT.raft": ["numpy", "torch"], "modelscope.models.cv.video_stabilization.DUT.RAFT.corr": ["alt_cuda_corr", "torch"], "modelscope.models.cv.video_stabilization.DUT.DUT_raft": ["numpy", "cv2", "sys", "torch"], "modelscope.models.cv.video_stabilization.DUT.config": ["__future__", "easydict"], "modelscope.models.cv.video_stabilization.DUT.rf_det_so": ["torch"], "modelscope.models.cv.video_stabilization.DUT.rf_det_module": ["torch"], "modelscope.models.cv.video_stabilization.DUT.MotionPro": ["numpy", "math", "cv2", "torch", "os"], "modelscope.models.cv.video_stabilization.DUT.Smoother": ["numpy", "math", "torch"], "modelscope.models.cv.video_stabilization.DUTRAFTStabilizer": ["numpy", "math", "typing", "cv2", "torch", "tempfile", "sys", "os"], "modelscope.models.cv.video_stabilization.utils.math_utils": ["numpy", "torch"], "modelscope.models.cv.video_stabilization.utils.MedianFilter": ["numpy", "math", "cv2", "torch"], "modelscope.models.cv.video_stabilization.utils.WarpUtils": ["numpy", "tqdm", "torch"], "modelscope.models.cv.video_stabilization.utils.ProjectionUtils": ["numpy", "math", "cv2", "torch"], "modelscope.models.cv.video_stabilization.utils.IterativeSmooth": ["numpy", "math", "os", "torch"], "modelscope.models.cv.video_stabilization.utils.RAFTUtils": ["numpy", "scipy", "torch"], "modelscope.models.cv.video_stabilization.utils.image_utils": ["skimage", "torch"], "modelscope.models.cv.facial_landmark_confidence.flc.manual_landmark_net": ["math", "torch"], "modelscope.models.cv.facial_landmark_confidence.flc.facial_landmark_confidence": ["numpy", "cv2", "PIL", "torch", "os"], "modelscope.models.multi_modal.videocomposer.annotator.histogram.palette": ["skimage", "numpy", "os", "sklearn"], "modelscope.models.multi_modal.videocomposer.annotator.sketch.sketch_simplification": ["math", "os", "torch"], "modelscope.models.multi_modal.videocomposer.annotator.sketch.pidinet": ["math", "os", "torch"], "modelscope.models.multi_modal.videocomposer.annotator.util": ["numpy", "os", "cv2"], "modelscope.models.multi_modal.videocomposer.unet_sd": ["fairscale", "functools", "math", "torch", "config", "einops", "rotary_embedding_torch", "os"], "modelscope.models.multi_modal.videocomposer.autoencoder": ["numpy", "torch"], "modelscope.models.multi_modal.videocomposer.models.clip": ["math", "os", "torch"], "modelscope.models.multi_modal.videocomposer.models.midas": ["math", "os", "torch"], "modelscope.models.multi_modal.videocomposer.diffusion": ["math", "torch"], "modelscope.models.multi_modal.videocomposer.ops.losses": ["math", "torch"], "modelscope.models.multi_modal.videocomposer.ops.utils": ["numpy", "gzip", "skvideo", "math", "binascii", "PIL", "urllib", "sys", "einops", "zipfile", "hashlib", "requests", "io", "time", "oss2", "torchvision", "pickle", "json", "torch", "base64", "copy", "multiprocessing", "logging", "os", "imageio", "glob"], "modelscope.models.multi_modal.videocomposer.ops.degration": ["numpy", "torchvision", "math", "torch", "datetime", "os", "random", "scipy"], "modelscope.models.multi_modal.videocomposer.ops.random_mask": ["numpy", "cv2"], "modelscope.models.multi_modal.videocomposer.ops.distributed": ["numpy", "collections", "functools", "pickle", "torch"], "modelscope.models.multi_modal.videocomposer.videocomposer_model": ["typing", "torch", "einops", "os", "copy", "pynvml", "open_clip"], "modelscope.models.multi_modal.videocomposer.config": ["easydict", "datetime", "logging", "os", "torch"], "modelscope.models.multi_modal.videocomposer.dpm_solver": ["math", "torch"], "modelscope.models.multi_modal.videocomposer.utils.utils": ["numpy", "gzip", "skvideo", "math", "binascii", "PIL", "urllib", "sys", "einops", "zipfile", "hashlib", "requests", "random", "io", "time", "oss2", "torchvision", "pickle", "json", "torch", "base64", "copy", "multiprocessing", "logging", "os", "imageio", "glob"], "modelscope.models.multi_modal.videocomposer.utils.config": ["yaml", "json", "argparse", "copy", "os"], "modelscope.models.multi_modal.videocomposer.utils.distributed": ["functools", "logging", "torch", "pickle"], "modelscope.models.multi_modal.videocomposer.data.samplers": ["numpy", "json", "os", "torch"], "modelscope.models.multi_modal.videocomposer.data.transforms": ["PIL", "numpy", "torchvision", "math", "random", "torch"], "modelscope.models.multi_modal.videocomposer.data.tokenizers": ["regex", "html", "gzip", "functools", "torch", "ftfy", "os", "tokenizers"], "modelscope.models.multi_modal.videocomposer.clip": ["numpy", "open_clip", "torchvision", "torch"], "modelscope.models.multi_modal.videocomposer.mha_flash": ["numpy", "time", "flash_attn", "random", "math", "os", "torch"], "modelscope.models.multi_modal.ofa_for_all_tasks": ["functools", "re", "math", "json", "string", "typing", "torch", "os"], "modelscope.models.multi_modal.team.utils": ["numpy", "collections", "transformers", "typing", "torch"], "modelscope.models.multi_modal.team.team_model": ["numpy", "torchvision", "typing", "torch", "PIL", "cv2", "tokenizers"], "modelscope.models.multi_modal.stable_diffusion.stable_diffusion_xl": ["transformers", "torchvision", "functools", "packaging", "typing", "torch", "os", "diffusers", "random"], "modelscope.models.multi_modal.stable_diffusion.stable_diffusion": ["transformers", "functools", "packaging", "typing", "torch", "diffusers", "os"], "modelscope.models.multi_modal.mmr.models.module_cross": ["__future__", "collections", "json", "torch", "logging"], "modelscope.models.multi_modal.mmr.models.modeling": ["collections", "torch", "platform", "os", "types"], "modelscope.models.multi_modal.mmr.models.until_module": ["numpy", "math", "logging", "torch"], "modelscope.models.multi_modal.mmr.models.dynamic_inverted_softmax": ["numpy"], "modelscope.models.multi_modal.mmr.models.module_clip": ["collections", "tqdm", "warnings", "torch", "typing", "urllib", "hashlib", "os"], "modelscope.models.multi_modal.mmr.models.clip_for_mm_video_embedding": ["numpy", "decord", "json", "typing", "torch", "PIL", "urllib", "uuid", "tempfile", "os", "random"], "modelscope.models.multi_modal.mmr.models.tokenization_clip": ["regex", "html", "gzip", "ftfy", "functools", "os"], "modelscope.models.multi_modal.mmr.dataloaders.rawvideo_util": ["numpy", "torchvision", "cv2", "PIL", "torch"], "modelscope.models.multi_modal.clip.bert_tokenizer": ["__future__", "collections", "unicodedata", "re", "os", "six"], "modelscope.models.multi_modal.clip.configuration_bert": ["__future__", "logging"], "modelscope.models.multi_modal.clip.model": ["numpy", "collections", "json", "typing", "torch", "os"], "modelscope.models.multi_modal.clip.modeling_bert": ["__future__", "math", "json", "torch", "sys", "logging", "os", "io"], "modelscope.models.multi_modal.gemm.gemm_base": ["numpy", "collections", "json", "typing", "torch", "os"], "modelscope.models.multi_modal.gemm.tokenizer": ["regex", "html", "gzip", "ftfy", "functools", "os", "torch"], "modelscope.models.multi_modal.gemm.gemm_model": ["numpy", "torchvision", "json", "typing", "torch", "PIL", "os"], "modelscope.models.multi_modal.mplug_owl.modeling_mplug_owl": ["transformers", "dataclasses", "math", "typing", "torch", "os", "copy", "logging", "random", "io"], "modelscope.models.multi_modal.mplug_owl.configuration_mplug_owl": ["transformers", "os", "typing", "copy"], "modelscope.models.multi_modal.video_synthesis.unet_sd": ["math", "einops", "torch"], "modelscope.models.multi_modal.video_synthesis.autoencoder": ["numpy", "torch"], "modelscope.models.multi_modal.video_synthesis.diffusion": ["torch"], "modelscope.models.multi_modal.video_synthesis.text_to_video_synthesis_model": ["typing", "torch", "einops", "os", "open_clip"], "modelscope.models.multi_modal.prost.models.module_cross": ["__future__", "collections", "math", "json", "shutil", "torch", "tarfile", "tempfile", "copy", "logging", "os"], "modelscope.models.multi_modal.prost.models.modeling": ["collections", "torch", "platform", "os", "types"], "modelscope.models.multi_modal.prost.models.until_module": ["numpy", "math", "torch", "copy", "logging"], "modelscope.models.multi_modal.prost.models.module_clip": ["collections", "tqdm", "warnings", "torch", "typing", "urllib", "hashlib", "os"], "modelscope.models.multi_modal.prost.models.prost_model": ["numpy", "decord", "json", "typing", "torch", "PIL", "urllib", "uuid", "tempfile", "os", "random"], "modelscope.models.multi_modal.prost.models.tokenization_clip": ["regex", "html", "gzip", "ftfy", "functools", "os"], "modelscope.models.multi_modal.prost.models.until_config": ["__future__", "json", "shutil", "torch", "tarfile", "tempfile", "copy", "logging", "os"], "modelscope.models.multi_modal.prost.dataloaders.rawvideo_util": ["numpy", "torchvision", "cv2", "PIL", "torch"], "modelscope.models.multi_modal.efficient_diffusion_tuning.sd_lora": ["dataclasses", "diffusers", "os", "typing", "torch"], "modelscope.models.multi_modal.efficient_diffusion_tuning.control_sd_lora": ["dataclasses", "typing", "torch", "diffusers", "os"], "modelscope.models.multi_modal.efficient_diffusion_tuning.efficient_stable_diffusion": ["transformers", "functools", "typing", "torch", "diffusers", "os", "swift"], "modelscope.models.multi_modal.mgeo.token_classification": ["torch"], "modelscope.models.multi_modal.mgeo.backbone": ["transformers", "dataclasses", "math", "warnings", "torch", "typing", "os", "random"], "modelscope.models.multi_modal.mgeo.text_ranking": ["torch"], "modelscope.models.multi_modal.mgeo.text_classification": ["torch"], "modelscope.models.multi_modal.vldoc.conv_fpn_trans": ["collections", "apex", "timm", "torch", "random"], "modelscope.models.multi_modal.vldoc.modeling_layout_roberta": ["transformers", "math", "packaging", "os", "torch"], "modelscope.models.multi_modal.vldoc.tokenization": ["transformers", "os"], "modelscope.models.multi_modal.vldoc.convnext": ["timm", "os", "torch"], "modelscope.models.multi_modal.vldoc.processing": ["numpy", "collections", "torchvision", "timm", "typing", "torch", "PIL", "cv2"], "modelscope.models.multi_modal.vldoc.model": ["torchvision", "re", "math", "json", "torch", "sys", "copy", "logging", "os"], "modelscope.models.multi_modal.vldoc.transformer_local": ["copy", "torch"], "modelscope.models.multi_modal.ofa.tokenization_ofa": ["collections", "transformers", "os", "typing"], "modelscope.models.multi_modal.ofa.configuration_ofa": ["transformers", "warnings"], "modelscope.models.multi_modal.ofa.tokenization_ofa_fast": ["json", "transformers", "typing", "tokenizers"], "modelscope.models.multi_modal.ofa.utils.utils": ["typing", "torch"], "modelscope.models.multi_modal.ofa.utils.constant": [], "modelscope.models.multi_modal.ofa.vit": ["collections", "fairseq", "torch"], "modelscope.models.multi_modal.ofa.resnet": ["torch"], "modelscope.models.multi_modal.ofa.configuration_mmspeech": ["transformers", "warnings"], "modelscope.models.multi_modal.ofa.modeling_mmspeech": ["numpy", "fairseq", "transformers", "apex", "dataclasses", "math", "packaging", "typing", "torch"], "modelscope.models.multi_modal.ofa.generate.incremental_decoding_utils": ["uuid", "typing", "torch"], "modelscope.models.multi_modal.ofa.generate.search": ["math", "typing", "torch"], "modelscope.models.multi_modal.ofa.generate.multihead_attention": ["math", "fairseq", "typing", "torch"], "modelscope.models.multi_modal.ofa.generate.sequence_generator": ["math", "sys", "typing", "torch"], "modelscope.models.multi_modal.ofa.generate.ngram_repeat_block": ["fairseq", "math", "warnings", "typing", "torch"], "modelscope.models.multi_modal.ofa.generate.utils": ["collections", "amp_C", "itertools", "torch_xla", "torch"], "modelscope.models.multi_modal.ofa.generate.token_generation_constraints": ["collections", "typing", "torch"], "modelscope.models.multi_modal.ofa.modeling_ofa": ["transformers", "apex", "dataclasses", "math", "packaging", "typing", "torch", "random"], "modelscope.models.multi_modal.dpm_solver_pytorch": ["math", "torch"], "modelscope.models.multi_modal.rleg.rleg": ["torchvision", "typing", "torch"], "modelscope.models.multi_modal.rleg.model": ["json", "os", "torch"], "modelscope.models.multi_modal.mplug.clip.clip": ["collections", "typing", "torch"], "modelscope.models.multi_modal.mplug.modeling_mplug": ["transformers", "math", "typing", "torch", "os"], "modelscope.models.multi_modal.mplug.configuration_mplug": ["yaml", "transformers", "os", "typing"], "modelscope.models.multi_modal.mplug.predictor": ["__future__", "torch"], "modelscope.models.multi_modal.mplug.mvit": ["numpy", "collections", "fairscale", "functools", "timm", "torch"], "modelscope.models.multi_modal.guided_diffusion.gaussian_diffusion": ["enum", "math", "numpy", "torch"], "modelscope.models.multi_modal.guided_diffusion.respace": ["numpy", "torch"], "modelscope.models.multi_modal.guided_diffusion.script": [], "modelscope.models.multi_modal.guided_diffusion.unet": ["numpy", "transformers", "abc", "math", "torch"], "modelscope.models.multi_modal.mplug_for_all_tasks": ["os", "typing"], "modelscope.models.multi_modal.clip_interrogator.model": ["numpy", "transformers", "math", "PIL", "hashlib", "requests", "open_clip", "time", "torchvision", "dataclasses", "tqdm", "typing", "torch", "safetensors", "os"], "modelscope.models.multi_modal.diffusion.diffusion": ["math", "torch"], "modelscope.models.multi_modal.diffusion.unet_upsampler_256": ["functools", "math", "torch"], "modelscope.models.multi_modal.diffusion.unet_upsampler_1024": ["math", "torch"], "modelscope.models.multi_modal.diffusion.tokenizer": ["unicodedata", "collections", "__future__", "six"], "modelscope.models.multi_modal.diffusion.model": ["numpy", "json", "typing", "torch", "os"], "modelscope.models.multi_modal.diffusion.structbert": ["numpy", "__future__", "copy", "math", "json", "six", "torch"], "modelscope.models.multi_modal.diffusion.unet_generator": ["math", "torch"], "modelscope.models.multi_modal.image_to_video.image_to_video_model": ["typing", "torch", "os", "copy", "random"], "modelscope.models.multi_modal.image_to_video.modules.autoencoder": ["numpy", "collections", "torch"], "modelscope.models.multi_modal.image_to_video.modules.embedder": ["numpy", "torchvision", "os", "open_clip", "torch"], "modelscope.models.multi_modal.image_to_video.modules.unet_i2v": ["xformers", "einops", "fairscale", "rotary_embedding_torch", "math", "os", "torch"], "modelscope.models.multi_modal.image_to_video.utils.transforms": ["PIL", "numpy", "torchvision", "math", "random", "torch"], "modelscope.models.multi_modal.image_to_video.utils.seed": ["numpy", "random", "torch"], "modelscope.models.multi_modal.image_to_video.utils.diffusion": ["math", "torch"], "modelscope.models.multi_modal.image_to_video.utils.shedule": ["math", "torch"], "modelscope.models.multi_modal.image_to_video.utils.config": ["easydict", "datetime", "logging", "os", "torch"], "modelscope.models.multi_modal.multi_stage_diffusion.gaussian_diffusion": ["math", "torch"], "modelscope.models.multi_modal.multi_stage_diffusion.upsampler": ["math", "torch"], "modelscope.models.multi_modal.multi_stage_diffusion.xglm": ["math", "torch"], "modelscope.models.multi_modal.multi_stage_diffusion.tokenizer": ["html", "regex", "gzip", "transformers", "ftfy", "functools", "torch"], "modelscope.models.multi_modal.multi_stage_diffusion.model": ["numpy", "math", "json", "typing", "torch", "PIL", "os"], "modelscope.models.multi_modal.multi_stage_diffusion.prior": ["math", "torch"], "modelscope.models.multi_modal.multi_stage_diffusion.decoder": ["math", "torch"], "modelscope.models.multi_modal.multi_stage_diffusion.clip": ["math", "torch"], "modelscope.models.multi_modal.soonet.utils": ["decord", "tqdm", "numpy", "copy"], "modelscope.models.multi_modal.soonet.blocks": ["math", "torch"], "modelscope.models.multi_modal.soonet.swin_transformer": ["numpy", "torch"], "modelscope.models.multi_modal.soonet.tokenizer": ["html", "regex", "gzip", "ftfy", "functools", "torch"], "modelscope.models.multi_modal.soonet.model": ["os", "torch"], "modelscope.models.multi_modal.soonet.clip": ["numpy", "collections", "warnings", "typing", "torch"], "modelscope.models.multi_modal.ofa_for_text_to_image_synthesis_model": ["numpy", "torchvision", "json", "typing", "torch", "PIL", "taming", "pkg_resources", "os"], "modelscope.models.multi_modal.video_to_video.modules.unet_v2v": ["xformers", "einops", "fairscale", "rotary_embedding_torch", "math", "os", "torch"], "modelscope.models.multi_modal.video_to_video.modules.autoencoder": ["numpy", "collections", "torch"], "modelscope.models.multi_modal.video_to_video.modules.embedder": ["numpy", "torchvision", "os", "open_clip", "torch"], "modelscope.models.multi_modal.video_to_video.utils.transforms": ["PIL", "numpy", "torchvision", "math", "random", "torch"], "modelscope.models.multi_modal.video_to_video.utils.seed": ["numpy", "random", "torch"], "modelscope.models.multi_modal.video_to_video.utils.config": ["easydict", "datetime", "logging", "os", "torch"], "modelscope.models.multi_modal.video_to_video.utils.schedules_sdedit": ["math", "torch"], "modelscope.models.multi_modal.video_to_video.utils.solvers_sdedit": ["torchsde", "tqdm", "torch"], "modelscope.models.multi_modal.video_to_video.utils.diffusion_sdedit": ["random", "torch"], "modelscope.models.multi_modal.video_to_video.video_to_video_model": ["typing", "torch", "os", "copy", "random"], "modelscope.models.builder": [], "modelscope.metrics.video_super_resolution_metric.metric_util": ["numpy"], "modelscope.metrics.video_super_resolution_metric.niqe": ["numpy", "math", "scipy", "cv2"], "modelscope.metrics.video_super_resolution_metric.matlab_functions": ["numpy", "math", "torch"], "modelscope.metrics.video_super_resolution_metric.video_super_resolution_metric": ["numpy", "typing"], "modelscope.metrics.ned_metric": ["numpy", "typing"], "modelscope.metrics.image_portrait_enhancement_metric": ["numpy", "typing", "cv2"], "modelscope.metrics.image_quality_assessment_degradation_metric": ["numpy", "collections", "tqdm", "typing", "cv2", "torch", "tempfile", "sys", "os", "scipy"], "modelscope.metrics.base": ["abc", "typing"], "modelscope.metrics.ciderD.ciderD_scorer": ["numpy", "__future__", "collections", "math", "pdb", "copy", "os", "six"], "modelscope.metrics.ciderD.ciderD": ["__future__"], "modelscope.metrics.prediction_saving_wrapper": ["typing"], "modelscope.metrics.text_ranking_metric": ["numpy", "typing"], "modelscope.metrics.video_summarization_metric": ["numpy", "typing"], "modelscope.metrics.map_metric": ["numpy", "typing"], "modelscope.metrics.ocr_recognition_metric": ["numpy", "torch", "typing", "edit_distance"], "modelscope.metrics.bleu_metric": ["sacrebleu", "itertools", "typing"], "modelscope.metrics.image_inpainting_metric": ["numpy", "scipy", "typing", "torch"], "modelscope.metrics.inbatch_recall_metric": ["numpy", "typing", "torch"], "modelscope.metrics.token_classification_metric": ["numpy", "importlib", "typing"], "modelscope.metrics.loss_metric": ["numpy", "sklearn", "typing"], "modelscope.metrics.sequence_classification_metric": ["numpy", "sklearn", "typing"], "modelscope.metrics.text_generation_metric": ["nltk", "contextlib", "typing", "sys", "rouge"], "modelscope.metrics.image_instance_segmentation_metric": ["numpy", "collections", "typing", "pycocotools", "tempfile", "os"], "modelscope.metrics.image_denoise_metric": ["numpy", "cv2", "typing", "torch"], "modelscope.metrics.audio_noise_metric": ["typing"], "modelscope.metrics.accuracy_metric": ["numpy", "typing"], "modelscope.metrics.translation_evaluation_metric": ["pandas", "importlib", "typing"], "modelscope.metrics.image_color_enhance_metric": ["numpy", "typing", "cv2"], "modelscope.metrics.movie_scene_segmentation_metric": ["numpy", "typing"], "modelscope.metrics.video_stabilization_metric": ["numpy", "tqdm", "typing", "cv2", "tempfile", "sys", "os"], "modelscope.metrics.ppl_metric": ["numpy", "math", "typing", "torch"], "modelscope.metrics.builder": ["typing"], "modelscope.metrics.action_detection_evaluator": ["numpy", "collections", "pandas", "copy", "detectron2", "logging", "os", "scipy"], "modelscope.metrics.image_quality_assessment_mos_metric": ["numpy", "tqdm", "typing", "torch", "cv2", "tempfile", "sys", "os", "scipy"], "modelscope.metrics.video_frame_interpolation_metric": ["numpy", "lpips", "math", "typing", "torch"], "modelscope.metrics.image_colorization_metric": ["numpy", "torchvision", "typing", "cv2", "torch", "scipy"], "modelscope.metrics.referring_video_object_segmentation_metric": ["numpy", "tqdm", "typing", "torch", "pycocotools"], "modelscope.pipelines.base": ["numpy", "abc", "functools", "packaging", "typing", "torch", "threading", "os", "multiprocessing", "random"], "modelscope.pipelines.pipeline_template": ["numpy", "typing"], "modelscope.pipelines.nlp.document_segmentation_pipeline": ["numpy", "re", "typing", "torch", "datasets"], "modelscope.pipelines.nlp.user_satisfaction_estimation_pipeline": ["numpy", "typing", "torch"], "modelscope.pipelines.nlp.summarization_pipeline": ["typing", "torch"], "modelscope.pipelines.nlp.document_grounded_dialog_retrieval_pipeline": ["numpy", "json", "typing", "faiss", "os"], "modelscope.pipelines.nlp.word_segmentation_pipeline": ["typing", "torch"], "modelscope.pipelines.nlp.distributed_gpt_moe_pipeline": ["typing", "torch"], "modelscope.pipelines.nlp.codegeex_code_translation_pipeline": ["typing"], "modelscope.pipelines.nlp.document_grounded_dialog_rerank_pipeline": ["numpy", "time", "pprint", "transformers", "collections", "re", "ujson", "typing", "torch", "sys", "os", "random"], "modelscope.pipelines.nlp.language_identification_pipline": ["numpy", "re", "typing", "os", "tensorflow"], "modelscope.pipelines.nlp.extractive_summarization_pipeline": ["numpy", "re", "typing", "torch", "datasets"], "modelscope.pipelines.nlp.zero_shot_classification_pipeline": ["typing", "scipy", "torch"], "modelscope.pipelines.nlp.table_question_answering_pipeline": ["transformers", "json", "typing", "torch", "os"], "modelscope.pipelines.nlp.named_entity_recognition_pipeline": ["typing"], "modelscope.pipelines.nlp.word_alignment_pipeline": ["numpy", "typing"], "modelscope.pipelines.nlp.faq_question_answering_pipeline": ["typing"], "modelscope.pipelines.nlp.text_classification_pipeline": ["numpy", "typing", "torch"], "modelscope.pipelines.nlp.distributed_plug_pipeline": ["typing", "torch"], "modelscope.pipelines.nlp.text_generation_pipeline": ["transformers", "os", "typing", "torch"], "modelscope.pipelines.nlp.dialog_modeling_pipeline": ["typing"], "modelscope.pipelines.nlp.dialog_state_tracking_pipeline": ["typing"], "modelscope.pipelines.nlp.automatic_post_editing_pipeline": ["html", "numpy", "sentencepiece", "typing", "sacremoses", "jieba", "os", "tensorflow"], "modelscope.pipelines.nlp.dialog_intent_prediction_pipeline": ["typing"], "modelscope.pipelines.nlp.distributed_gpt3_pipeline": ["typing", "torch"], "modelscope.pipelines.nlp.sentence_embedding_pipeline": ["numpy", "typing", "torch"], "modelscope.pipelines.nlp.glm130b_text_generation_pipeline": ["typing"], "modelscope.pipelines.nlp.fill_mask_pipeline": ["numpy", "typing"], "modelscope.pipelines.nlp.document_grounded_dialog_generate_pipeline": ["typing"], "modelscope.pipelines.nlp.machine_reading_comprehension_pipeline": ["typing", "torch"], "modelscope.pipelines.nlp.text_ranking_pipeline": ["numpy", "typing"], "modelscope.pipelines.nlp.text_error_correction_pipeline": ["typing", "torch"], "modelscope.pipelines.nlp.fasttext_text_classification_pipeline": ["numpy", "sentencepiece", "typing", "fasttext", "os"], "modelscope.pipelines.nlp.polylm_text_generation_pipeline": ["os", "typing", "torch"], "modelscope.pipelines.nlp.codegeex_code_generation_pipeline": ["typing"], "modelscope.pipelines.nlp.translation_evaluation_pipeline": ["numpy", "enum", "typing", "torch", "os"], "modelscope.pipelines.nlp.interactive_translation_pipeline": ["numpy", "typing", "sacremoses", "subword_nmt", "jieba", "os", "tensorflow"], "modelscope.pipelines.nlp.token_classification_pipeline": ["numpy", "typing", "torch"], "modelscope.pipelines.nlp.siamese_uie_pipeline": ["time", "math", "tqdm", "json", "typing", "torch", "pathlib", "copy", "logging", "os", "scipy"], "modelscope.pipelines.nlp.translation_quality_estimation_pipeline": ["transformers", "typing", "torch", "os", "io"], "modelscope.pipelines.nlp.canmt_translation_pipeline": ["typing", "os", "sacremoses", "torch"], "modelscope.pipelines.nlp.conversational_text_to_sql_pipeline": ["typing", "text2sql_lgesql", "torch"], "modelscope.pipelines.nlp.mglm_text_summarization_pipeline": ["os", "typing"], "modelscope.pipelines.nlp.translation_pipeline": ["numpy", "typing", "sacremoses", "subword_nmt", "jieba", "os", "tensorflow"], "modelscope.pipelines.nlp.fid_dialogue_pipeline": ["re", "typing", "torch"], "modelscope.pipelines.nlp.information_extraction_pipeline": ["typing", "torch"], "modelscope.pipelines.nlp.feature_extraction_pipeline": ["os", "typing", "torch"], "modelscope.pipelines.science.protein_structure_pipeline": ["numpy", "time", "json", "typing", "torch", "os", "unicore"], "modelscope.pipelines.util": ["os", "typing"], "modelscope.pipelines.audio.language_recognition_eres2net_pipeline": ["numpy", "typing", "torch", "soundfile", "os", "io", "torchaudio"], "modelscope.pipelines.audio.asr_inference_pipeline": ["yaml", "json", "os", "typing"], "modelscope.pipelines.audio.ans_dfsmn_pipeline": ["numpy", "collections", "typing", "librosa", "torch", "sys", "soundfile", "os", "io"], "modelscope.pipelines.audio.speaker_verification_light_pipeline": ["numpy", "typing", "torch", "soundfile", "os", "io", "torchaudio"], "modelscope.pipelines.audio.speaker_verification_eres2net_pipeline": ["soundfile", "io", "typing", "torch"], "modelscope.pipelines.audio.timestamp_pipeline": ["yaml", "json", "typing", "funasr", "os"], "modelscope.pipelines.audio.text_to_speech_pipeline": ["numpy", "typing"], "modelscope.pipelines.audio.speaker_diarization_dialogue_detection_pipeline": ["numpy", "typing"], "modelscope.pipelines.audio.lm_infer_pipeline": ["os", "typing"], "modelscope.pipelines.audio.speaker_verification_pipeline": ["yaml", "shutil", "os", "typing"], "modelscope.pipelines.audio.speaker_change_locating_pipeline": ["numpy", "typing", "torch", "soundfile", "io", "torchaudio"], "modelscope.pipelines.audio.segmentation_clustering_pipeline": ["numpy", "ast", "typing", "torch", "soundfile", "io", "torchaudio"], "modelscope.pipelines.audio.speaker_diarization_semantic_speaker_turn_detection_pipeline": ["numpy", "typing", "torch"], "modelscope.pipelines.audio.linear_aec_pipeline": ["numpy", "yaml", "typing", "torch", "importlib", "os", "scipy"], "modelscope.pipelines.audio.kws_kwsbp_pipeline": ["typing", "json", "os"], "modelscope.pipelines.audio.kws_farfield_pipeline": ["wave", "numpy", "typing", "soundfile", "io"], "modelscope.pipelines.audio.language_recognition_pipeline": ["numpy", "typing", "torch", "soundfile", "os", "io", "torchaudio"], "modelscope.pipelines.audio.speech_separation_pipeline": ["yaml", "json", "typing", "funasr", "os"], "modelscope.pipelines.audio.asr_wenet_inference_pipeline": ["typing"], "modelscope.pipelines.audio.punctuation_processing_pipeline": ["yaml", "shutil", "os", "typing"], "modelscope.pipelines.audio.ans_pipeline": ["numpy", "typing", "librosa", "torch", "soundfile", "io"], "modelscope.pipelines.audio.inverse_text_processing_pipeline": ["yaml", "shutil", "os", "typing"], "modelscope.pipelines.audio.separation_pipeline": ["numpy", "typing", "torch", "soundfile", "io"], "modelscope.pipelines.audio.speaker_diarization_pipeline": ["numpy", "yaml", "json", "typing", "shutil", "os"], "modelscope.pipelines.audio.voice_activity_detection_pipeline": ["yaml", "json", "typing", "funasr", "os"], "modelscope.pipelines.audio.speaker_verification_rdino_pipeline": ["soundfile", "io", "typing", "torch"], "modelscope.pipelines.cv.image_detection_pipeline": ["numpy", "typing"], "modelscope.pipelines.cv.face_recognition_onnx_fm_pipeline": ["numpy", "onnxruntime", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.ocr_utils.model_dla34": ["numpy", "math", "os", "torch"], "modelscope.pipelines.cv.ocr_utils.table_process": ["numpy", "copy", "math", "torch", "random", "cv2"], "modelscope.pipelines.cv.ocr_utils.resnet18_v1": ["tf_slim", "tensorflow"], "modelscope.pipelines.cv.ocr_utils.model_vlpt": ["math", "os", "sys", "torch"], "modelscope.pipelines.cv.ocr_utils.utils": ["pyclipper", "numpy", "shapely", "cv2"], "modelscope.pipelines.cv.ocr_utils.model_resnet18_half": ["os", "torch"], "modelscope.pipelines.cv.ocr_utils.model_convnext_transformer": ["torch"], "modelscope.pipelines.cv.ocr_utils.resnet_utils": ["collections", "tf_slim", "tensorflow"], "modelscope.pipelines.cv.ocr_utils.ops": ["numpy", "absl", "math", "shutil", "cv2", "uuid", "sys", "os", "tensorflow"], "modelscope.pipelines.cv.ocr_utils.model_resnet_mutex_v4_linewithchar": ["tf_slim", "tensorflow"], "modelscope.pipelines.cv.ocr_utils.ocr_modules.timm_tinyc": ["collections", "itertools", "copy", "functools", "math", "logging", "torch"], "modelscope.pipelines.cv.ocr_utils.ocr_modules.convnext": ["torch"], "modelscope.pipelines.cv.ocr_utils.ocr_modules.vitstr": ["__future__", "functools", "torch", "copy", "logging"], "modelscope.pipelines.cv.product_segmentation_pipeline": ["numpy", "typing"], "modelscope.pipelines.cv.pedestrian_attribute_recognition_pipeline": ["numpy", "torchvision", "json", "typing", "torch", "cv2", "PIL", "os"], "modelscope.pipelines.cv.image_quality_assessment_degradation_pipeline": ["numpy", "torchvision", "math", "typing", "torch", "cv2", "tempfile"], "modelscope.pipelines.cv.video_frame_interpolation_pipeline": ["numpy", "torchvision", "subprocess", "math", "typing", "cv2", "torch", "tempfile", "os", "glob"], "modelscope.pipelines.cv.tbs_detection_utils.utils": ["numpy", "__future__", "torchvision", "pandas", "torch", "matplotlib", "PIL", "colorsys", "os"], "modelscope.pipelines.cv.image_quality_assessment_man_pipeline": ["numpy", "torchvision", "math", "typing", "torch", "cv2", "tempfile"], "modelscope.pipelines.cv.face_quality_assessment_pipeline": ["numpy", "onnxruntime", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.object_detection_3d_pipeline": ["numpy", "typing", "torch", "PIL", "cv2", "tempfile", "os"], "modelscope.pipelines.cv.video_instance_segmentation_pipeline": ["numpy", "tqdm", "typing", "torch", "cv2", "mmcv", "os"], "modelscope.pipelines.cv.face_attribute_recognition_pipeline": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.video_depth_estimation_pipeline": ["typing"], "modelscope.pipelines.cv.image_skychange_pipeline": ["numpy", "time", "typing", "cv2", "PIL", "pdb"], "modelscope.pipelines.cv.hicossl_video_embedding_pipeline": ["math", "os", "typing", "torch"], "modelscope.pipelines.cv.video_single_object_tracking_pipeline": ["os", "typing", "cv2"], "modelscope.pipelines.cv.card_detection_pipeline": ["typing"], "modelscope.pipelines.cv.vision_efficient_tuning_pipeline": ["numpy", "torchvision", "typing", "torch"], "modelscope.pipelines.cv.video_object_segmentation_pipeline": ["numpy", "torchvision", "typing", "torch", "PIL", "os"], "modelscope.pipelines.cv.hand_static_pipeline": ["numpy", "typing"], "modelscope.pipelines.cv.ddcolor_image_colorization_pipeline": ["numpy", "torchvision", "typing", "cv2", "torch"], "modelscope.pipelines.cv.face_recognition_pipeline": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.image_style_transfer_pipeline": ["numpy", "os", "typing", "cv2"], "modelscope.pipelines.cv.body_3d_keypoints_pipeline": ["numpy", "mpl_toolkits", "typing", "cv2", "torch", "tempfile", "datetime", "os", "matplotlib"], "modelscope.pipelines.cv.image_human_parsing_pipeline": ["numpy", "torchvision", "typing", "torch"], "modelscope.pipelines.cv.image_cartoon_pipeline": ["numpy", "typing", "cv2", "os", "tensorflow"], "modelscope.pipelines.cv.image_bts_depth_estimation_pipeline": ["numpy", "typing", "cv2", "torch", "albumentations"], "modelscope.pipelines.cv.mobile_image_super_resolution_pipeline": ["numpy", "torchvision", "typing", "torch", "skimage"], "modelscope.pipelines.cv.realtime_video_object_detection_pipeline": ["numpy", "torchvision", "json", "typing", "torch", "cv2", "PIL", "os"], "modelscope.pipelines.cv.general_recognition_pipeline": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "os"], "modelscope.pipelines.cv.image_try_on_pipeline": ["numpy", "typing", "torch"], "modelscope.pipelines.cv.content_check_pipeline": ["numpy", "torchvision", "typing", "torch", "PIL", "cv2", "os"], "modelscope.pipelines.cv.nerf_recon_acc_pipeline": ["typing"], "modelscope.pipelines.cv.ocr_recognition_pipeline": [], "modelscope.pipelines.cv.action_detection_pipeline": ["math", "os", "typing"], "modelscope.pipelines.cv.animal_recognition_pipeline": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "os"], "modelscope.pipelines.cv.face_liveness_xc_pipeline": ["numpy", "onnxruntime", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.human_reconstruction_pipeline": ["numpy", "typing", "shutil", "torch", "os", "trimesh"], "modelscope.pipelines.cv.image_salient_detection_pipeline": ["typing"], "modelscope.pipelines.cv.table_recognition_pipeline": ["numpy", "math", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.ddpm_semantic_segmentation_pipeline": ["torchvision", "typing", "torch"], "modelscope.pipelines.cv.vop_retrieval_pipeline": ["numpy", "gzip", "collections", "math", "pickle", "tqdm", "typing", "torch", "os", "random"], "modelscope.pipelines.cv.skin_retouching_pipeline": ["numpy", "torchvision", "typing", "cv2", "PIL", "torch", "os", "tensorflow"], "modelscope.pipelines.cv.image_portrait_enhancement_pipeline": ["numpy", "math", "typing", "torch", "PIL", "cv2", "scipy"], "modelscope.pipelines.cv.image_to_image_translation_pipeline": ["numpy", "torchvision", "typing", "cv2", "torch", "PIL", "sys", "os", "io"], "modelscope.pipelines.cv.face_human_hand_detection_pipeline": ["numpy", "typing"], "modelscope.pipelines.cv.image_inpainting_sdv2_pipeline": ["numpy", "math", "typing", "cv2", "torch", "tempfile", "sys", "diffusers", "os"], "modelscope.pipelines.cv.image_matting_pipeline": ["numpy", "typing", "cv2", "os", "tensorflow"], "modelscope.pipelines.cv.image_depth_estimation_pipeline": ["numpy", "typing", "cv2", "PIL", "torch"], "modelscope.pipelines.cv.tinynas_classification_pipeline": ["torchvision", "math", "typing", "torch", "os"], "modelscope.pipelines.cv.image_super_resolution_pasd_pipeline": ["numpy", "transformers", "torchvision", "typing", "torch", "PIL", "tempfile", "diffusers", "os"], "modelscope.pipelines.cv.pointcloud_sceneflow_estimation_pipeline": ["numpy", "plyfile", "typing", "torch"], "modelscope.pipelines.cv.image_paintbyexample_pipeline": ["numpy", "torchvision", "typing", "torch", "PIL", "cv2", "einops"], "modelscope.pipelines.cv.face_recognition_onnx_ir_pipeline": ["numpy", "onnxruntime", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.image_classification_pipeline": ["numpy", "typing", "torch"], "modelscope.pipelines.cv.image_to_image_generate_pipeline": ["numpy", "torchvision", "typing", "torch", "PIL", "cv2", "os"], "modelscope.pipelines.cv.face_reconstruction_pipeline": ["numpy", "shutil", "cv2", "PIL", "torch", "tensorflow", "face_alignment", "typing", "scipy", "os", "io"], "modelscope.pipelines.cv.movie_scene_segmentation_pipeline": ["typing", "torch"], "modelscope.pipelines.cv.vop_retrieval_se_pipeline": ["numpy", "gzip", "typing", "torch", "os"], "modelscope.pipelines.cv.indoor_layout_estimation_pipeline": ["numpy", "typing", "cv2"], "modelscope.pipelines.cv.image_driving_perception_pipeline": ["numpy", "os", "typing", "cv2"], "modelscope.pipelines.cv.panorama_depth_estimation_pipeline": ["numpy", "typing", "cv2", "PIL", "torch"], "modelscope.pipelines.cv.video_inpainting_pipeline": ["typing"], "modelscope.pipelines.cv.text_driven_segmentation_pipleline": ["typing"], "modelscope.pipelines.cv.image_colorization_pipeline": ["numpy", "torchvision", "typing", "torch", "PIL", "cv2"], "modelscope.pipelines.cv.nerf_recon_4k_pipeline": ["typing"], "modelscope.pipelines.cv.action_recognition_pipeline": ["math", "os", "typing", "torch"], "modelscope.pipelines.cv.image_instance_segmentation_pipeline": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.language_guided_video_summarization_pipeline": ["numpy", "shutil", "cv2", "torch", "typing", "PIL", "tempfile", "os", "clip", "random"], "modelscope.pipelines.cv.nerf_recon_vq_compression_pipeline": ["typing"], "modelscope.pipelines.cv.shop_segmentation_pipleline": ["typing"], "modelscope.pipelines.cv.face_processing_base_pipeline": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.facial_expression_recognition_pipeline": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.image_deblur_pipeline": ["torchvision", "typing", "torch"], "modelscope.pipelines.cv.video_panoptic_segmentation_pipeline": ["numpy", "tqdm", "typing", "torch", "cv2", "mmcv", "os"], "modelscope.pipelines.cv.body_2d_keypoints_pipeline": ["numpy", "torchvision", "json", "typing", "torch", "cv2", "PIL", "os"], "modelscope.pipelines.cv.video_stabilization_pipeline": ["numpy", "subprocess", "math", "typing", "cv2", "torch", "tempfile", "os", "glob"], "modelscope.pipelines.cv.image_open_vocabulary_detection_pipeline": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.tinynas_detection_pipeline": ["typing"], "modelscope.pipelines.cv.video_super_resolution_pipeline": ["numpy", "torchvision", "subprocess", "math", "typing", "cv2", "torch", "tempfile", "os"], "modelscope.pipelines.cv.arc_face_recognition_pipeline": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.maskdino_instance_segmentation_pipeline": ["torchvision", "typing", "torch"], "modelscope.pipelines.cv.video_colorization_pipeline": ["numpy", "torchvision", "subprocess", "typing", "cv2", "PIL", "torch", "tempfile", "os"], "modelscope.pipelines.cv.face_emotion_pipeline": ["numpy", "typing"], "modelscope.pipelines.cv.tbs_detection_pipeline": ["numpy", "typing", "torch", "cv2", "PIL", "colorsys", "os"], "modelscope.pipelines.cv.image_restoration_pipeline": ["typing"], "modelscope.pipelines.cv.image_debanding_pipeline": ["torchvision", "typing", "torch"], "modelscope.pipelines.cv.image_editing_pipeline": ["numpy", "torchvision", "tqdm", "typing", "torch", "PIL", "diffusers", "os"], "modelscope.pipelines.cv.image_super_resolution_pipeline": ["numpy", "typing", "cv2", "PIL", "torch"], "modelscope.pipelines.cv.cmdssl_video_embedding_pipeline": ["numpy", "torchvision", "decord", "typing", "torch", "PIL", "os"], "modelscope.pipelines.cv.video_multi_object_tracking_pipeline": ["os", "typing", "torch"], "modelscope.pipelines.cv.image_face_fusion_pipeline": ["numpy", "typing"], "modelscope.pipelines.cv.face_liveness_ir_pipeline": ["numpy", "onnxruntime", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.image_denoise_pipeline": ["torchvision", "typing", "torch"], "modelscope.pipelines.cv.panorama_depth_estimation_s2net_pipeline": ["numpy", "typing", "cv2", "PIL", "torch"], "modelscope.pipelines.cv.image_reid_person_pipeline": ["torchvision", "math", "typing", "torch", "PIL", "os"], "modelscope.pipelines.cv.image_body_reshaping_pipeline": ["typing"], "modelscope.pipelines.cv.mtcnn_face_detection_pipeline": ["os", "typing", "torch"], "modelscope.pipelines.cv.face_detection_pipeline": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.ocr_detection_pipeline": ["numpy", "math", "typing", "torch", "cv2", "os", "tf_slim", "tensorflow"], "modelscope.pipelines.cv.video_deinterlace_pipeline": ["numpy", "torchvision", "subprocess", "math", "typing", "cv2", "torch", "tempfile", "os"], "modelscope.pipelines.cv.product_retrieval_embedding_pipeline": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "os"], "modelscope.pipelines.cv.license_plate_detection_pipeline": ["numpy", "math", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.video_category_pipeline": ["numpy", "torchvision", "decord", "json", "typing", "torch", "PIL", "os"], "modelscope.pipelines.cv.image_color_enhance_pipeline": ["torchvision", "typing", "torch"], "modelscope.pipelines.cv.referring_video_object_segmentation_pipeline": ["numpy", "torchvision", "tqdm", "typing", "torch", "moviepy", "PIL", "tempfile", "einops"], "modelscope.pipelines.cv.virtual_try_on_pipeline": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.crowd_counting_pipeline": ["numpy", "torchvision", "math", "typing", "torch", "PIL"], "modelscope.pipelines.cv.vidt_pipeline": ["torchvision", "typing", "torch"], "modelscope.pipelines.cv.image_panoptic_segmentation_pipeline": ["numpy", "typing", "cv2", "PIL", "torch"], "modelscope.pipelines.cv.image_mvs_depth_estimation_pipeline": ["shutil", "tempfile", "os", "typing"], "modelscope.pipelines.cv.bad_image_detecting_pipeline": ["numpy", "typing", "torch"], "modelscope.pipelines.cv.text_to_360panorama_image_pipeline": ["numpy", "typing", "torch", "basicsr", "PIL", "diffusers", "random", "realesrgan"], "modelscope.pipelines.cv.image_matching_pipeline": ["numpy", "typing", "cv2", "PIL", "torch"], "modelscope.pipelines.cv.image_structured_model_probing_pipeline": ["numpy", "torchvision", "math", "typing", "torch", "mmcv", "os"], "modelscope.pipelines.cv.face_recognition_ood_pipeline": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.human_image_generation_pipeline": ["numpy", "typing", "torch"], "modelscope.pipelines.cv.video_summarization_pipeline": ["numpy", "tqdm", "typing", "cv2", "torch", "os"], "modelscope.pipelines.cv.video_human_matting_pipeline": ["numpy", "moviepy", "cv2", "typing", "torch", "os"], "modelscope.pipelines.cv.fast_instance_segmentation_pipeline": ["numpy", "torchvision", "typing", "torch"], "modelscope.pipelines.cv.image_defrcn_fewshot_pipeline": ["numpy", "os", "typing", "torch"], "modelscope.pipelines.cv.facial_landmark_confidence_pipeline": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.lineless_table_recognition_pipeline": ["numpy", "math", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.surface_recon_common_pipeline": ["typing"], "modelscope.pipelines.cv.image_quality_assessment_mos_pipeline": ["numpy", "torchvision", "math", "typing", "torch", "cv2", "tempfile"], "modelscope.pipelines.cv.ulfd_face_detection_pipeline": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.controllable_image_generation_pipeline": ["numpy", "subprocess", "math", "typing", "cv2", "torch", "tempfile", "os", "glob"], "modelscope.pipelines.cv.image_semantic_segmentation_pipeline": ["numpy", "typing", "cv2", "PIL", "torch"], "modelscope.pipelines.cv.image_inpainting_pipeline": ["numpy", "typing", "cv2", "PIL", "torch"], "modelscope.pipelines.cv.vision_middleware_pipeline": ["numpy", "torchvision", "math", "typing", "torch", "mmcv", "os"], "modelscope.pipelines.cv.retina_face_detection_pipeline": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.live_category_pipeline": ["numpy", "torchvision", "decord", "typing", "torch", "PIL", "os"], "modelscope.pipelines.cv.face_image_generation_pipeline": ["numpy", "typing", "cv2", "PIL", "torch", "os"], "modelscope.pipelines.cv.mask_face_recognition_pipeline": ["numpy", "collections", "typing", "torch", "PIL", "cv2", "os"], "modelscope.pipelines.cv.motion_generation_pipeline": ["numpy", "typing", "torch", "tempfile", "os"], "modelscope.pipelines.cv.mog_face_detection_pipeline": ["numpy", "os", "typing"], "modelscope.pipelines.multi_modal.prost_text_video_retrieval_pipeline": ["typing"], "modelscope.pipelines.multi_modal.asr_pipeline": ["typing", "torch"], "modelscope.pipelines.multi_modal.document_vl_embedding_pipeline": ["typing", "torch"], "modelscope.pipelines.multi_modal.efficient_diffusion_tuning_pipeline": ["numpy", "torchvision", "typing", "cv2", "torch", "PIL"], "modelscope.pipelines.multi_modal.video_multi_modal_embedding_pipeline": ["typing"], "modelscope.pipelines.multi_modal.soonet_video_temporal_grounding_pipeline": ["numpy", "torchvision", "typing", "torch", "os"], "modelscope.pipelines.multi_modal.image_captioning_pipeline": ["numpy", "typing", "torch"], "modelscope.pipelines.multi_modal.video_question_answering_pipeline": ["typing", "torch"], "modelscope.pipelines.multi_modal.ocr_recognition_pipeline": ["typing", "torch"], "modelscope.pipelines.multi_modal.image_text_retrieval_pipeline": ["typing", "torch"], "modelscope.pipelines.multi_modal.cone2_pipeline.cones2_inference_pipeline": ["numpy", "math", "tqdm", "typing", "torch", "cv2", "PIL", "diffusers"], "modelscope.pipelines.multi_modal.visual_entailment_pipeline": ["typing", "torch"], "modelscope.pipelines.multi_modal.image_to_video_pipeline": ["subprocess", "typing", "torch", "cv2", "tempfile", "einops", "os"], "modelscope.pipelines.multi_modal.sudoku_pipeline": ["typing", "torch"], "modelscope.pipelines.multi_modal.visual_question_answering_pipeline": ["typing", "torch"], "modelscope.pipelines.multi_modal.team_multi_modal_similarity_pipeline": ["typing"], "modelscope.pipelines.multi_modal.text2sql_pipeline": ["typing", "torch"], "modelscope.pipelines.multi_modal.video_captioning_pipeline": ["typing", "torch"], "modelscope.pipelines.multi_modal.mgeo_ranking_pipeline": ["numpy", "typing", "torch"], "modelscope.pipelines.multi_modal.multi_modal_embedding_pipeline": ["typing"], "modelscope.pipelines.multi_modal.videocomposer_pipeline": ["numpy", "time", "torchvision", "functools", "subprocess", "mvextractor", "typing", "cv2", "torch", "PIL", "tempfile", "os", "random", "imageio"], "modelscope.pipelines.multi_modal.multimodal_dialogue_pipeline": ["typing", "torch"], "modelscope.pipelines.multi_modal.disco_guided_diffusion_pipeline.disco_guided_diffusion": ["numpy", "gc", "torchvision", "math", "json", "cv2", "torch", "PIL", "importlib", "clip", "os"], "modelscope.pipelines.multi_modal.disco_guided_diffusion_pipeline.utils": ["numpy", "fractions", "math", "warnings", "torch"], "modelscope.pipelines.multi_modal.generative_multi_modal_embedding_pipeline": ["typing"], "modelscope.pipelines.multi_modal.diffusers_wrapped.devices": ["modules", "sys", "contextlib", "torch"], "modelscope.pipelines.multi_modal.diffusers_wrapped.stable_diffusion.stable_diffusion_pipeline": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "diffusers", "os"], "modelscope.pipelines.multi_modal.diffusers_wrapped.stable_diffusion.chinese_stable_diffusion_pipeline": ["numpy", "transformers", "typing", "torch", "cv2", "PIL", "diffusers"], "modelscope.pipelines.multi_modal.diffusers_wrapped.pasd_pipeline": ["numpy", "transformers", "torchvision", "warnings", "inspect", "PIL", "torch", "typing", "diffusers", "os"], "modelscope.pipelines.multi_modal.diffusers_wrapped.diffusers_pipeline": ["os", "typing"], "modelscope.pipelines.multi_modal.diffusers_wrapped.vaehook": ["time", "gc", "math", "tqdm", "torch"], "modelscope.pipelines.multi_modal.video_to_video_pipeline": ["subprocess", "typing", "torch", "cv2", "tempfile", "einops", "os"], "modelscope.pipelines.multi_modal.text_to_video_synthesis_pipeline": ["typing", "cv2", "torch", "tempfile", "einops", "os"], "modelscope.pipelines.multi_modal.visual_grounding_pipeline": ["typing", "torch"], "modelscope.pipelines.multi_modal.text_to_image_synthesis_pipeline": ["typing", "torch"], "modelscope.pipelines.multi_modal.gridvlp_pipeline": ["traceback", "numpy", "time", "transformers", "json", "typing", "torch", "PIL", "os"], "modelscope.pipelines.builder": ["os", "typing"], "modelscope.preprocessors.tts": ["kantts", "os", "typing"], "modelscope.preprocessors.multi_modal": ["numpy", "torchvision", "re", "decord", "json", "timm", "typing", "torch", "PIL", "os", "io"], "modelscope.preprocessors.base": ["abc", "os", "typing"], "modelscope.preprocessors.audio": ["numpy", "typing", "torch", "scipy", "os", "io"], "modelscope.preprocessors.nlp.text_generation_preprocessor": ["numpy", "os", "typing", "torch"], "modelscope.preprocessors.nlp.zero_shot_classification_preprocessor": ["typing"], "modelscope.preprocessors.nlp.token_classification_thai_preprocessor": ["typing"], "modelscope.preprocessors.nlp.relation_extraction_preprocessor": ["transformers", "typing"], "modelscope.preprocessors.nlp.fill_mask_preprocessor": ["numpy", "abc", "re", "typing", "torch", "os"], "modelscope.preprocessors.nlp.text_clean": ["codecs", "re", "sys"], "modelscope.preprocessors.nlp.space_T_en.fields.process_dataset": ["pickle", "os", "sys", "text2sql_lgesql"], "modelscope.preprocessors.nlp.space_T_en.fields.preprocess_dataset": ["text2sql_lgesql"], "modelscope.preprocessors.nlp.space_T_en.fields.parse": [], "modelscope.preprocessors.nlp.space_T_en.fields.common_utils": ["nltk", "numpy", "sqlite3", "itertools", "text2sql_lgesql", "os"], "modelscope.preprocessors.nlp.space_T_en.conversational_text_to_sql_preprocessor": ["json", "typing", "torch", "text2sql_lgesql", "os"], "modelscope.preprocessors.nlp.transformers_tokenizer": ["json", "collections", "transformers", "os"], "modelscope.preprocessors.nlp.word_alignment_preprocessor": ["numpy", "typing", "torch", "itertools", "os"], "modelscope.preprocessors.nlp.space_T_cn.fields.struct": [], "modelscope.preprocessors.nlp.space_T_cn.fields.schema_link": ["re"], "modelscope.preprocessors.nlp.space_T_cn.fields.database": ["tqdm", "json", "sqlite3"], "modelscope.preprocessors.nlp.space_T_cn.table_question_answering_preprocessor": ["transformers", "os", "typing", "torch"], "modelscope.preprocessors.nlp.space.args": ["json", "argparse"], "modelscope.preprocessors.nlp.space.preprocess": ["os", "glob"], "modelscope.preprocessors.nlp.space.data_loader": ["numpy", "math", "os"], "modelscope.preprocessors.nlp.space.fields.intent_field": ["numpy", "time", "collections", "re", "tqdm", "json", "itertools", "os", "multiprocessing", "random", "glob"], "modelscope.preprocessors.nlp.space.fields.gen_field": ["numpy", "collections", "json", "asyncio", "itertools", "os", "random"], "modelscope.preprocessors.nlp.space.tensorlistdataset": ["torch"], "modelscope.preprocessors.nlp.space.dialog_state_tracking_preprocessor": ["typing"], "modelscope.preprocessors.nlp.space.tokenizer": ["regex", "__future__", "collections", "functools", "json", "unicodedata", "sys", "logging", "os"], "modelscope.preprocessors.nlp.space.dialog_modeling_preprocessor": ["os", "typing"], "modelscope.preprocessors.nlp.space.batch": [], "modelscope.preprocessors.nlp.space.dialog_intent_prediction_preprocessor": ["typing", "json", "os"], "modelscope.preprocessors.nlp.space.dst_processors": ["numpy", "logging", "re", "tqdm", "json", "six"], "modelscope.preprocessors.nlp.space.sampler": ["numpy"], "modelscope.preprocessors.nlp.space.lazy_dataset": ["json"], "modelscope.preprocessors.nlp.document_grounded_dialog_retrieval_preprocessor": ["transformers", "os", "typing", "torch"], "modelscope.preprocessors.nlp.document_segmentation_preprocessor": ["typing"], "modelscope.preprocessors.nlp.dialog_classification_use_preprocessor": ["transformers", "typing", "torch"], "modelscope.preprocessors.nlp.canmt_translation": ["typing", "sacremoses", "torch", "subword_nmt", "jieba", "os"], "modelscope.preprocessors.nlp.document_grounded_dialog_rerank_preprocessor": ["transformers", "typing", "torch", "copy", "os"], "modelscope.preprocessors.nlp.text_classification_preprocessor": ["numpy", "typing"], "modelscope.preprocessors.nlp.utils": ["numpy", "collections", "transformers", "json", "typing", "os"], "modelscope.preprocessors.nlp.bert_seq_cls_tokenizer": ["transformers", "typing"], "modelscope.preprocessors.nlp.token_classification_preprocessor": ["numpy", "typing", "torch"], "modelscope.preprocessors.nlp.machine_reading_comprehension_preprocessor": ["transformers", "os", "torch"], "modelscope.preprocessors.nlp.faq_question_answering_preprocessor": ["typing", "torch"], "modelscope.preprocessors.nlp.text_ranking_preprocessor": ["transformers", "typing"], "modelscope.preprocessors.nlp.token_classification_viet_preprocessor": ["typing", "torch"], "modelscope.preprocessors.nlp.sentence_embedding_preprocessor": ["typing"], "modelscope.preprocessors.nlp.feature_extraction_preprocessor": ["numpy", "typing"], "modelscope.preprocessors.nlp.mgeo_ranking_preprocessor": ["transformers", "typing", "torch"], "modelscope.preprocessors.nlp.text_error_correction": ["transformers", "os", "typing", "torch"], "modelscope.preprocessors.nlp.translation_evaluation_preprocessor": ["transformers", "typing", "torch"], "modelscope.preprocessors.nlp.mglm_summarization_preprocessor": ["re", "os", "typing"], "modelscope.preprocessors.nlp.siamese_uie_preprocessor": ["transformers", "typing"], "modelscope.preprocessors.nlp.document_grounded_dialog_generate_preprocessor": ["transformers", "os", "typing", "torch"], "modelscope.preprocessors.image": ["numpy", "typing", "cv2", "PIL", "io"], "modelscope.preprocessors.ofa.sudoku": ["numpy", "typing", "torch"], "modelscope.preprocessors.ofa.base": ["numpy", "re", "json", "string", "torch", "PIL", "os", "io", "torchaudio"], "modelscope.preprocessors.ofa.image_classification": ["torchvision", "functools", "timm", "typing", "torch", "PIL"], "modelscope.preprocessors.ofa.text_to_image_synthesis": ["typing", "torch"], "modelscope.preprocessors.ofa.text2sql": ["re", "typing", "torch", "os", "random"], "modelscope.preprocessors.ofa.visual_grounding": ["numpy", "torchvision", "typing", "torch", "PIL"], "modelscope.preprocessors.ofa.ocr_recognition": ["torchvision", "unicodedata2", "zhconv", "typing", "torch"], "modelscope.preprocessors.ofa.utils.random_help": ["torch_xla", "torch"], "modelscope.preprocessors.ofa.utils.transforms": ["PIL", "numpy", "torchvision", "random", "torch"], "modelscope.preprocessors.ofa.utils.collate": ["numpy", "typing", "torch"], "modelscope.preprocessors.ofa.utils.get_tables": ["traceback", "sys", "sqlite3"], "modelscope.preprocessors.ofa.utils.audio_helper": ["numpy", "typing", "torch"], "modelscope.preprocessors.ofa.utils.vision_helper": ["numpy", "cv2"], "modelscope.preprocessors.ofa.utils.constant": [], "modelscope.preprocessors.ofa.utils.text2phone": [], "modelscope.preprocessors.ofa.utils.bridge_content_encoder": ["functools", "difflib", "sqlite3", "typing", "rapidfuzz"], "modelscope.preprocessors.ofa.asr": ["fairseq", "typing", "librosa", "torch", "pathlib", "random", "soundfile", "os"], "modelscope.preprocessors.ofa.visual_entailment": ["PIL", "torchvision", "typing", "torch"], "modelscope.preprocessors.ofa.text_classification": ["typing", "torch"], "modelscope.preprocessors.ofa.summarization": ["typing", "torch"], "modelscope.preprocessors.ofa.image_captioning": ["torchvision", "typing", "torch"], "modelscope.preprocessors.ofa.visual_question_answering": ["PIL", "torchvision", "typing", "torch"], "modelscope.preprocessors.speaker": ["typing", "torch"], "modelscope.preprocessors.kws": ["yaml", "os", "typing"], "modelscope.preprocessors.movie_scene_segmentation.transforms": ["numpy", "torchvision", "numbers", "typing", "torch", "PIL", "os", "random"], "modelscope.preprocessors.asr": ["os", "typing"], "modelscope.preprocessors.science.uni_fold": ["numpy", "gzip", "ipdb", "tarfile", "hashlib", "requests", "random", "time", "unittest", "re", "pickle", "json", "tqdm", "typing", "torch", "pathlib", "logging", "os"], "modelscope.preprocessors.common": ["numpy", "collections", "time", "typing", "torch"], "modelscope.preprocessors.cv.controllable_image_generation": ["numpy", "torchvision", "math", "typing", "torch", "cv2", "PIL", "os"], "modelscope.preprocessors.cv.video_super_resolution": ["collections", "os", "cv2"], "modelscope.preprocessors.cv.action_detection_mapper": ["numpy", "copy", "detectron2", "decord", "random", "scipy", "torch"], "modelscope.preprocessors.cv.bad_image_detecting_preprocessor": ["numpy", "torchvision", "math", "typing", "torch", "PIL"], "modelscope.preprocessors.cv.cv2_transforms": ["numpy", "collections", "math", "torch", "random", "numbers", "cv2"], "modelscope.preprocessors.cv.image_classification_preprocessor": ["numpy", "torchvision", "typing", "torch", "cv2", "PIL", "os"], "modelscope.preprocessors.cv.image_quality_assessment_man": ["numpy", "torchvision", "math", "typing", "torch", "PIL"], "modelscope.preprocessors.cv.image_quality_assessment_mos": ["numpy", "torchvision", "math", "typing", "cv2"], "modelscope.preprocessors.cv.mmcls_preprocessor": ["numpy", "os", "typing"], "modelscope.preprocessors.cv.timer": ["time"], "modelscope.preprocessors.cv.util": ["collections", "sys", "os", "shutil"], "modelscope.preprocessors.cv.video_stabilization": ["numpy", "cv2", "torch"], "modelscope.preprocessors.cv.image_restoration_preprocessor": ["numpy", "torchvision", "math", "typing", "torch", "PIL"], "modelscope.preprocessors.video": ["numpy", "torchvision", "decord", "math", "torch", "urllib", "uuid", "tempfile", "os", "random"], "modelscope.preprocessors.builder": [], "modelscope.trainers.parallel.utils": [], "modelscope.trainers.parallel.builder": ["torch"], "modelscope.trainers.lrscheduler.warmup.base": ["torch"], "modelscope.trainers.lrscheduler.warmup.warmup": [], "modelscope.trainers.lrscheduler.builder": ["packaging", "inspect", "torch"], "modelscope.trainers.base": ["abc", "time", "os", "typing"], "modelscope.trainers.hooks.distributed.megatron_hook": ["megatron_util", "os", "shutil", "torch"], "modelscope.trainers.hooks.distributed.deepspeed_hook": ["transformers", "deepspeed", "functools", "math", "megatron_util", "shutil", "torch", "os"], "modelscope.trainers.hooks.distributed.ddp_hook": [], "modelscope.trainers.hooks.hook": ["functools"], "modelscope.trainers.hooks.iter_timer_hook": ["time"], "modelscope.trainers.hooks.early_stop_hook": ["numpy"], "modelscope.trainers.hooks.evaluation_hook": ["collections", "typing"], "modelscope.trainers.hooks.checkpoint.checkpoint_processor": ["re", "os", "shutil"], "modelscope.trainers.hooks.checkpoint.load_checkpoint_hook": ["numpy", "packaging", "typing", "torch", "random"], "modelscope.trainers.hooks.checkpoint.checkpoint_hook": ["numpy", "json", "shutil", "torch", "typing", "os", "random"], "modelscope.trainers.hooks.compression.sparsity_hook": ["os"], "modelscope.trainers.hooks.compression.utils": ["torch"], "modelscope.trainers.hooks.optimizer.apex_optimizer_hook": ["packaging", "logging", "torch"], "modelscope.trainers.hooks.optimizer.base": ["logging", "torch"], "modelscope.trainers.hooks.optimizer.torch_optimizer_hook": ["logging"], "modelscope.trainers.hooks.lr_scheduler_hook": [], "modelscope.trainers.hooks.clip_clamp_logit_scale_hook": ["torch"], "modelscope.trainers.hooks.builder": [], "modelscope.trainers.hooks.priority": ["enum", "typing"], "modelscope.trainers.hooks.logger.base": ["numpy", "numbers", "abc", "torch"], "modelscope.trainers.hooks.logger.text_logger_hook": ["collections", "json", "torch", "datetime", "os"], "modelscope.trainers.hooks.logger.tensorboard_hook": ["numpy", "os", "torch"], "modelscope.trainers.nlp_trainer": ["numpy", "os", "typing", "torch"], "modelscope.trainers.nlp.faq_question_answering_trainer": ["numpy", "collections", "dataclasses", "functools", "contextlib", "torch", "typing", "distutils"], "modelscope.trainers.nlp.siamese_uie_trainer": ["numpy", "collections", "time", "math", "json", "typing", "torch", "random", "os"], "modelscope.trainers.nlp.plug_trainer": ["deepspeed", "megatron_util", "typing", "torch", "os"], "modelscope.trainers.nlp.document_grounded_dialog_generate_trainer": ["collections", "transformers", "re", "tqdm", "json", "sacrebleu", "string", "torch", "rouge", "os"], "modelscope.trainers.nlp.gpt_moe_trainer": ["collections", "megatron_util", "typing", "torch", "os"], "modelscope.trainers.nlp.document_grounded_dialog_retrieval_trainer": ["numpy", "transformers", "tqdm", "json", "torch", "faiss", "os"], "modelscope.trainers.nlp.space.dialog_modeling_trainer": ["numpy", "time", "os", "typing"], "modelscope.trainers.nlp.space.eval": ["nltk", "numpy", "collections", "math", "json", "sklearn"], "modelscope.trainers.nlp.space.trainer.gen_trainer": ["numpy", "time", "collections", "transformers", "tqdm", "json", "torch", "os"], "modelscope.trainers.nlp.space.trainer.intent_trainer": ["numpy", "time", "collections", "transformers", "tqdm", "json", "torch", "os"], "modelscope.trainers.nlp.space.dialog_intent_trainer": ["numpy", "os", "typing"], "modelscope.trainers.nlp.space.metrics.metrics_tracker": ["math", "collections"], "modelscope.trainers.nlp.text_ranking_trainer": ["numpy", "time", "dataclasses", "tqdm", "typing", "torch"], "modelscope.trainers.nlp.translation_evaluation_trainer": ["transformers", "math", "tqdm", "pandas", "typing", "torch", "os", "random"], "modelscope.trainers.nlp.table_question_answering_trainer": ["numpy", "time", "tqdm", "json", "typing", "torch", "os"], "modelscope.trainers.nlp.document_grounded_dialog_rerank_trainer": ["numpy", "time", "transformers", "typing", "torch", "os", "random"], "modelscope.trainers.nlp.gpt3_trainer": ["copy", "os", "typing", "torch"], "modelscope.trainers.nlp.text_generation_trainer": ["typing", "torch"], "modelscope.trainers.nlp.sequence_classification_trainer": ["numpy", "time", "typing"], "modelscope.trainers.nlp.csanmt_translation_trainer": ["tensorflow", "time", "os", "typing"], "modelscope.trainers.nlp.sentence_embedding_trainer": ["numpy", "time", "transformers", "dataclasses", "tqdm", "typing", "torch"], "modelscope.trainers.trainer": ["collections", "functools", "json", "inspect", "torch", "typing", "distutils", "copy", "os"], "modelscope.trainers.utils.log_buffer": ["numpy", "collections"], "modelscope.trainers.utils.inference": ["collections", "pickle", "tqdm", "shutil", "torch", "logging", "os"], "modelscope.trainers.training_args": ["re", "dataclasses", "addict", "json", "typing", "copy"], "modelscope.trainers.optimizer.child_tuning_adamw_optimizer": ["numpy", "math", "typing", "torch"], "modelscope.trainers.optimizer.builder": ["inspect", "typing", "torch"], "modelscope.trainers.audio.asr_trainer": ["json", "shutil", "typing", "tempfile", "funasr", "os"], "modelscope.trainers.audio.kws_nearfield_trainer": ["yaml", "re", "typing", "torch", "tensorboardX", "datetime", "copy", "os"], "modelscope.trainers.audio.tts_trainer": ["json", "typing", "shutil", "tempfile", "zipfile", "os"], "modelscope.trainers.audio.kws_farfield_trainer": ["numpy", "math", "pickle", "typing", "torch", "datetime", "os", "glob"], "modelscope.trainers.audio.kws_utils.det_utils": ["numpy", "json", "torch", "threading", "os", "kaldiio", "matplotlib", "glob"], "modelscope.trainers.audio.kws_utils.runtime_utils": ["collections", "stat", "codecs", "re", "json", "shutil", "sys", "os"], "modelscope.trainers.audio.kws_utils.model_utils": ["numpy", "yaml", "re", "shutil", "torch", "os", "glob"], "modelscope.trainers.audio.kws_utils.file_utils": ["re"], "modelscope.trainers.audio.kws_utils.batch_utils": ["numpy", "collections", "math", "typing", "torch", "sys", "datetime", "os"], "modelscope.trainers.audio.separation_trainer": ["numpy", "tqdm", "typing", "torch", "speechbrain", "csv", "os", "torchaudio"], "modelscope.trainers.audio.ans_trainer": [], "modelscope.trainers.cv.ocr_recognition_trainer": ["collections", "time", "torch"], "modelscope.trainers.cv.image_portrait_enhancement_trainer": ["collections", "torch"], "modelscope.trainers.cv.ocr_detection_db_trainer": ["numpy", "time", "math", "tqdm", "typing", "torch", "easydict", "datetime", "copy", "os"], "modelscope.trainers.cv.action_detection_trainer": ["typing", "torch", "fvcore", "detectron2", "os"], "modelscope.trainers.cv.image_detection_damoyolo_trainer": ["time", "math", "typing", "torch", "easydict", "datetime", "os"], "modelscope.trainers.cv.face_detection_scrfd_trainer": ["time", "os", "typing", "copy"], "modelscope.trainers.cv.image_instance_segmentation_trainer": [], "modelscope.trainers.cv.image_inpainting_trainer": ["collections", "time", "torch"], "modelscope.trainers.cv.nerf_recon_acc_trainer": ["numpy", "time", "tqdm", "typing", "cv2", "torch", "datetime", "os", "random", "glob"], "modelscope.trainers.cv.cartoon_translation_trainer": ["numpy", "packaging", "tqdm", "typing", "os", "tensorflow"], "modelscope.trainers.cv.image_defrcn_fewshot_detection_trainer": ["collections", "typing", "torch", "detectron2", "os"], "modelscope.trainers.cv.image_classifition_trainer": ["numpy", "time", "typing", "torch", "copy", "os"], "modelscope.trainers.cv.movie_scene_segmentation_trainer": [], "modelscope.trainers.cv.vision_efficient_tuning_trainer": ["typing", "torch"], "modelscope.trainers.cv.card_detection_scrfd_trainer": [], "modelscope.trainers.cv.referring_video_object_segmentation_trainer": ["os", "torch"], "modelscope.trainers.multi_modal.dreambooth_diffusion.dreambooth_diffusion_trainer": ["collections", "torchvision", "tqdm", "warnings", "shutil", "torch", "typing", "PIL", "itertools", "pathlib", "hashlib", "diffusers"], "modelscope.trainers.multi_modal.lora_diffusion.lora_diffusion_trainer": ["diffusers", "typing", "torch"], "modelscope.trainers.multi_modal.team.team_trainer_utils": ["PIL", "torchvision", "torch"], "modelscope.trainers.multi_modal.team.team_trainer": ["numpy", "collections", "typing", "torch", "sklearn", "os"], "modelscope.trainers.multi_modal.cones2.cones_trainer": ["torchvision", "torch", "PIL", "pathlib", "diffusers"], "modelscope.trainers.multi_modal.custom_diffusion.custom_diffusion_trainer": ["numpy", "torchvision", "tqdm", "json", "warnings", "torch", "typing", "PIL", "itertools", "pathlib", "os", "hashlib", "diffusers", "random"], "modelscope.trainers.multi_modal.stable_diffusion.stable_diffusion_trainer": ["typing", "torch"], "modelscope.trainers.multi_modal.clip.clip_trainer_utils": ["functools", "math", "inspect", "torch", "os"], "modelscope.trainers.multi_modal.clip.clip_trainer": ["math", "os", "typing", "torch"], "modelscope.trainers.multi_modal.lora_diffusion_xl.lora_diffusion_xl_trainer": ["diffusers", "typing", "torch"], "modelscope.trainers.multi_modal.efficient_diffusion_tuning.efficient_diffusion_tuning_trainer": ["typing", "torch"], "modelscope.trainers.multi_modal.ofa.ofa_trainer_utils": ["numpy", "transformers", "math", "os", "shutil", "torch"], "modelscope.trainers.multi_modal.ofa.ofa_trainer": ["functools", "math", "json", "shutil", "torch", "typing", "tempfile", "os"], "modelscope.trainers.multi_modal.mplug.mplug_trainer": ["collections", "typing", "torch"], "modelscope.trainers.multi_modal.mgeo_ranking_trainer": ["dataclasses", "typing", "torch"], "modelscope.trainers.builder": [], "modelscope.trainers.cli_argument_parser": ["dataclasses", "argparse", "typing"], "modelscope.trainers.default_config": ["typing"], "modelscope.msdatasets.task_datasets.sidd_image_denoising": [], "modelscope.msdatasets.task_datasets.reds_image_deblurring_dataset": [], "modelscope.msdatasets.task_datasets.video_summarization_dataset": [], "modelscope.msdatasets.task_datasets.torch_base_dataset": [], "modelscope.msdatasets.task_datasets.gopro_image_deblurring_dataset": [], "modelscope.msdatasets.download.download_manager": ["datasets"], "modelscope.msdatasets.download.dataset_builder": ["pandas", "typing", "datasets", "pyarrow", "os"], "modelscope.msdatasets.download.download_config": ["datasets", "typing"], "modelscope.msdatasets.data_loader.data_loader": ["abc", "datasets", "os", "typing"], "modelscope.msdatasets.data_loader.data_loader_manager": ["enum", "datasets", "os", "abc"], "modelscope.msdatasets.dataset_cls.dataset": ["math", "tqdm", "pandas", "datasets", "itertools", "copy", "os"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_colorization.image_colorization_dataset": ["numpy", "cv2", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.torch_custom_dataset": ["typing", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.language_guided_video_summarization_dataset": ["numpy", "h5py", "json", "torch", "os"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_quality_assessment_degradation.image_quality_assessment_degradation_dataset": ["torchvision"], "modelscope.msdatasets.dataset_cls.custom_datasets.bad_image_detecting.bad_image_detecting_dataset": [], "modelscope.msdatasets.dataset_cls.custom_datasets.video_super_resolution.video_super_resolution_dataset": ["numpy", "collections", "cv2", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_instance_segmentation_coco_dataset": ["numpy", "os", "pycocotools"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_recognition_dataset": ["numpy", "json", "cv2", "torch", "PIL", "lmdb", "os", "six"], "modelscope.msdatasets.dataset_cls.custom_datasets.video_frame_interpolation.data_utils": ["cv2", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.video_frame_interpolation.video_frame_interpolation_dataset": ["numpy", "cv2", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.reds_image_deblurring_dataset": ["numpy", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.evaluation.coco.coco_eval": ["collections", "os", "tempfile", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.build": ["copy", "math", "bisect", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.samplers.grouped_batch_sampler": ["itertools", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.samplers.iteration_based_batch_sampler": ["torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.samplers.distributed": ["math", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.transforms.build": [], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.transforms.transforms": ["numpy", "torchvision", "random", "cv2", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.datasets.coco": ["numpy", "cv2", "torchvision", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.datasets.mosaic_wrapper": ["numpy", "math", "cv2", "torch", "random"], "modelscope.msdatasets.dataset_cls.custom_datasets.damoyolo.collate_batch": [], "modelscope.msdatasets.dataset_cls.custom_datasets.video_summarization_dataset": ["numpy", "h5py", "json", "torch", "os"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_portrait_enhancement.data_utils": ["cv2", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_portrait_enhancement.image_portrait_enhancement_dataset": ["numpy", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_inpainting.aug": ["albumentations", "imgaug"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_inpainting.image_inpainting_dataset": ["numpy", "enum", "cv2", "albumentations", "os", "glob"], "modelscope.msdatasets.dataset_cls.custom_datasets.mgeo_ranking_dataset": ["random", "json", "typing", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.movie_scene_segmentation.movie_scene_segmentation_dataset": ["torchvision", "json", "torch", "copy", "random", "os"], "modelscope.msdatasets.dataset_cls.custom_datasets.movie_scene_segmentation.sampler": ["numpy", "random"], "modelscope.msdatasets.dataset_cls.custom_datasets.gopro_image_deblurring_dataset": ["numpy", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.easycv_base": ["os"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.measures.quad_measurer": ["numpy"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.measures.iou_evaluator": ["numpy", "collections", "shapely"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.data_loader": ["numpy", "math", "torch", "bisect", "imgaug"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.image_dataset": ["numpy", "functools", "math", "cv2", "torch", "bisect", "logging", "os", "glob"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.augmenter": ["imgaug"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.processes.data_process": [], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.processes.make_seg_detection_data": ["pyclipper", "numpy", "shapely", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.processes.normalize_image": ["numpy", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.processes.random_crop_data": ["numpy", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.processes.make_icdar_data": ["numpy", "collections", "cv2", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.processes.augment_data": ["numpy", "math", "imgaug", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.ocr_detection.processes.make_border_map": ["pyclipper", "numpy", "shapely", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.text_ranking_dataset": ["random", "typing", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.sidd_image_denoising.data_utils": ["cv2", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.sidd_image_denoising.transforms": ["random"], "modelscope.msdatasets.dataset_cls.custom_datasets.sidd_image_denoising.sidd_image_denoising_dataset": ["numpy", "cv2"], "modelscope.msdatasets.dataset_cls.custom_datasets.audio.kws_nearfield_processor": ["numpy", "json", "torch", "kaldiio", "random", "torchaudio"], "modelscope.msdatasets.dataset_cls.custom_datasets.audio.asr_dataset": ["os"], "modelscope.msdatasets.dataset_cls.custom_datasets.audio.kws_farfield_dataset": ["numpy", "math", "queue", "torch", "threading", "os"], "modelscope.msdatasets.dataset_cls.custom_datasets.audio.kws_nearfield_dataset": ["random", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.image_quality_assmessment_mos.image_quality_assessment_mos_dataset": [], "modelscope.msdatasets.dataset_cls.custom_datasets.veco_dataset": ["numpy", "datasets", "typing"], "modelscope.msdatasets.dataset_cls.custom_datasets.builder": [], "modelscope.msdatasets.dataset_cls.custom_datasets.referring_video_object_segmentation.referring_video_object_segmentation_dataset": ["numpy", "h5py", "torchvision", "tqdm", "json", "pandas", "torch", "pycocotools", "os", "glob"], "modelscope.msdatasets.dataset_cls.custom_datasets.referring_video_object_segmentation.transformers": ["PIL", "random", "torchvision", "torch"], "modelscope.msdatasets.dataset_cls.custom_datasets.video_stabilization.video_stabilization_dataset": [], "modelscope.msdatasets.context.dataset_context_config": ["typing"], "modelscope.msdatasets.data_files.data_files_manager": ["os", "typing", "datasets"], "modelscope.msdatasets.meta.data_meta_config": [], "modelscope.msdatasets.meta.data_meta_manager": ["collections", "json", "shutil", "datasets", "os"], "modelscope.msdatasets.utils.upload_utils": ["multiprocessing", "tqdm", "os"], "modelscope.msdatasets.utils.delete_utils": [], "modelscope.msdatasets.utils.dataset_utils": ["collections", "pandas", "os", "typing"], "modelscope.msdatasets.utils.oss_utils": ["__future__", "oss2", "datasets", "multiprocessing", "os"], "modelscope.msdatasets.utils.maxcompute_utils": ["math", "pandas"], "modelscope.msdatasets.audio.asr_dataset": [], "modelscope.msdatasets.auth.auth_config": ["http", "typing"], "modelscope.msdatasets.ms_dataset": ["numpy", "warnings", "typing", "datasets", "os"], "modelscope.exporters.torch_model_exporter": ["contextlib", "typing", "torch", "itertools", "os"], "modelscope.exporters.base": ["abc", "os", "typing"], "modelscope.exporters.nlp.csanmt_for_translation_exporter": ["tensorflow", "os", "typing"], "modelscope.exporters.nlp.sbert_for_zero_shot_classification_exporter": ["collections", "typing"], "modelscope.exporters.nlp.model_for_token_classification_exporter": ["collections", "typing", "torch"], "modelscope.exporters.nlp.sbert_for_sequence_classification_exporter": ["collections", "typing", "torch"], "modelscope.exporters.tf_model_exporter": ["tensorflow", "os", "typing"], "modelscope.exporters.audio.ans_dfsmn_exporter": ["os", "torch"], "modelscope.exporters.cv.object_detection_damoyolo_exporter": ["numpy", "functools", "typing", "torch", "onnx", "os"], "modelscope.exporters.cv.cartoon_translation_exporter": ["tensorflow", "packaging", "os", "typing"], "modelscope.exporters.cv.face_detection_scrfd_exporter": ["numpy", "functools", "typing", "torch", "onnx", "os"], "modelscope.exporters.multi_modal.stable_diffusion_exporter": ["collections", "packaging", "argparse", "shutil", "torch", "typing", "pathlib", "onnx", "diffusers", "os"], "modelscope.exporters.builder": []}, "version": "1.9.1", "md5": "986f04d5fd32f1972505a19cd9b80eed", "files_mtime": {"D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\aec\\layers\\activations.py": 1696497788.4881158, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\aec\\layers\\affine_transform.py": 1696497788.4891143, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\aec\\layers\\deep_fsmn.py": 1696497788.4901147, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\aec\\layers\\layer_base.py": 1696497788.4911182, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\aec\\layers\\uni_deep_fsmn.py": 1696497788.492118, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\aec\\network\\loss.py": 1696497788.4941163, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\aec\\network\\modulation_loss.py": 1696497788.4951158, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\aec\\network\\se_net.py": 1696497788.496117, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\ans\\complex_nn.py": 1696497788.498116, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\ans\\conv_stft.py": 1696497788.4991157, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\ans\\denoise_net.py": 1696497788.5001142, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\ans\\frcrn.py": 1696497788.5021152, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\ans\\layers\\activations.py": 1696497788.5061178, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\ans\\layers\\affine_transform.py": 1696497788.5071151, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\ans\\layers\\layer_base.py": 1696497788.5081167, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\ans\\layers\\uni_deep_fsmn.py": 1696497788.509116, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\ans\\se_module_complex.py": 1696497788.5031157, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\ans\\unet.py": 1696497788.5041177, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\asr\\generic_automatic_speech_recognition.py": 1696497788.511114, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\asr\\wenet_automatic_speech_recognition.py": 1696497788.511114, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\itn\\generic_inverse_text_processing.py": 1696497788.5131147, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\kws\\farfield\\fsmn.py": 1696497788.5171158, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\kws\\farfield\\fsmn_sele_v2.py": 1696497788.5181139, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\kws\\farfield\\fsmn_sele_v3.py": 1696497788.5191138, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\kws\\farfield\\model.py": 1696497788.5201163, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\kws\\farfield\\model_def.py": 1696497788.5211208, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\kws\\generic_key_word_spotting.py": 1696497788.5151136, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\kws\\nearfield\\cmvn.py": 1696497788.5221193, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\kws\\nearfield\\fsmn.py": 1696497788.523117, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\kws\\nearfield\\model.py": 1696497788.5241175, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\punc\\generic_punctuation.py": 1696497788.526117, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\separation\\layer_norm.py": 1696497788.5281174, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\separation\\m2\\conv_module.py": 1696497788.5321176, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\separation\\m2\\fsmn.py": 1696497788.533118, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\separation\\m2\\layer_norm.py": 1696497788.533118, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\separation\\m2\\mossformer.py": 1696497788.5341182, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\separation\\m2\\mossformer_block.py": 1696497788.5351183, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\separation\\mossformer.py": 1696497788.5291188, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\separation\\mossformer_block.py": 1696497788.5291188, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\separation\\mossformer_conv_module.py": 1696497788.5301178, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\sv\\cluster_backend.py": 1696497788.540115, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\sv\\DTDNN.py": 1696497788.5361156, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\sv\\DTDNN_layers.py": 1696497788.5371163, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\sv\\ecapa_tdnn.py": 1696497788.5411162, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\sv\\ERes2Net.py": 1696497788.5381148, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\sv\\ERes2Net_aug.py": 1696497788.539116, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\sv\\fusion.py": 1696497788.5421157, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\sv\\generic_speaker_verification.py": 1696497788.5431154, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\sv\\lanuage_recognition_eres2net.py": 1696497788.5441172, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\sv\\lanuage_recognition_model.py": 1696497788.5451167, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\sv\\pooling_layers.py": 1696497788.5461147, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\sv\\rdino.py": 1696497788.547115, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\sv\\speaker_change_locator.py": 1696497788.5481193, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\sv\\speaker_diarization_dialogue_detection.py": 1696497788.5491154, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\sv\\speaker_diarization_semantic_speaker_turn_detection.py": 1696497788.5501144, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\tts\\sambert_hifi.py": 1696497788.5521173, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\audio\\tts\\voice.py": 1696497788.5531166, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\base\\base_head.py": 1696497788.5551174, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\base\\base_model.py": 1696497788.5561154, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\base\\base_torch_head.py": 1696497788.55712, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\base\\base_torch_model.py": 1696497788.558118, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\builder.py": 1696497788.4851153, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\abnormal_object_detection\\mmdet_model.py": 1696497788.5611167, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\abnormal_object_detection\\mmdet_ms\\roi_head\\mask_scoring_roi_head.py": 1696497788.5641155, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\abnormal_object_detection\\mmdet_ms\\roi_head\\roi_extractors\\single_level_roi_extractor.py": 1696497788.5661151, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\action_detection\\action_detection_onnx.py": 1696497788.5681155, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\action_detection\\modules\\action_detection_pytorch.py": 1696497788.570114, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\action_detection\\modules\\resnet.py": 1696497788.5711148, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\action_recognition\\models.py": 1696497788.5741172, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\action_recognition\\s3dg.py": 1696497788.575116, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\action_recognition\\tada_convnext.py": 1696497788.5761156, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\action_recognition\\temporal_patch_shift_transformer.py": 1696497788.5771158, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\animal_recognition\\resnet.py": 1696497788.579115, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\animal_recognition\\splat.py": 1696497788.580115, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\bad_image_detecting\\bad_image_detecting.py": 1696497788.5821173, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\body_2d_keypoints\\hrnet_basic_modules.py": 1696497788.5841193, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\body_2d_keypoints\\hrnet_v2.py": 1696497788.5851192, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\body_2d_keypoints\\w48.py": 1696497788.5861194, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\body_3d_keypoints\\cannonical_pose\\body_3d_pose.py": 1696497788.5891178, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\body_3d_keypoints\\cannonical_pose\\canonical_pose_modules.py": 1696497788.5901175, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\body_3d_keypoints\\hdformer\\backbone.py": 1696497788.5921178, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\body_3d_keypoints\\hdformer\\block.py": 1696497788.594118, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\body_3d_keypoints\\hdformer\\directed_graph.py": 1696497788.5951214, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\body_3d_keypoints\\hdformer\\hdformer.py": 1696497788.597118, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\body_3d_keypoints\\hdformer\\hdformer_detector.py": 1696497788.5981164, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\body_3d_keypoints\\hdformer\\skeleton.py": 1696497788.5991204, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\cartoon\\facelib\\config.py": 1696497788.6051164, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\cartoon\\facelib\\facer.py": 1696497788.6081183, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\cartoon\\facelib\\face_detector.py": 1696497788.6061192, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\cartoon\\facelib\\face_landmark.py": 1696497788.6071193, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\cartoon\\facelib\\LK\\lk.py": 1696497788.6091213, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\cartoon\\loss.py": 1696497788.6011186, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\cartoon\\model_tf.py": 1696497788.6021185, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\cartoon\\mtcnn_pytorch\\src\\align_trans.py": 1696497788.612121, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\cartoon\\mtcnn_pytorch\\src\\matlab_cp2tform.py": 1696497788.6131206, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\cartoon\\network.py": 1696497788.603117, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\cartoon\\utils.py": 1696497788.603117, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\cmdssl_video_embedding\\c3d.py": 1696497788.6141198, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\cmdssl_video_embedding\\resnet2p1d.py": 1696497788.6151206, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\cmdssl_video_embedding\\resnet3d.py": 1696497788.617118, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\annotator.py": 1696497788.621115, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\midas\\api.py": 1696497788.6231189, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\midas\\midas\\base_model.py": 1696497788.6251187, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\midas\\midas\\blocks.py": 1696497788.626115, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\midas\\midas\\dpt_depth.py": 1696497788.6271176, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\midas\\midas\\midas_net.py": 1696497788.6271176, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\midas\\midas\\midas_net_custom.py": 1696497788.6281168, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\midas\\midas\\transforms.py": 1696497788.629119, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\midas\\midas\\vit.py": 1696497788.6301198, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\midas\\utils.py": 1696497788.6231189, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\mlsd\\mbv2_mlsd_large.py": 1696497788.6321168, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\mlsd\\utils.py": 1696497788.6331193, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\openpose\\body.py": 1696497788.6341197, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\openpose\\hand.py": 1696497788.6351197, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\openpose\\model.py": 1696497788.6361194, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\annotator\\openpose\\util.py": 1696497788.6371176, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\controllable_image_generation\\controlnet.py": 1696497788.6191168, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\crowd_counting\\cc_model.py": 1696497788.6381164, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\crowd_counting\\hrnet_aspp_relu.py": 1696497788.639117, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_attribute_recognition\\fair_face\\face_attribute_recognition.py": 1696497788.6421182, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\mogface\\models\\detectors.py": 1696497788.6451328, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\mogface\\models\\mogface.py": 1696497788.6461325, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\mogface\\models\\mogprednet.py": 1696497788.6461325, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\mogface\\models\\resnet.py": 1696497788.6471355, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\mogface\\models\\utils.py": 1696497788.6481173, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\mtcnn\\models\\box_utils.py": 1696497788.6511161, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\mtcnn\\models\\detector.py": 1696497788.6511161, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\mtcnn\\models\\first_stage.py": 1696497788.6521158, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\mtcnn\\models\\get_nets.py": 1696497788.6531324, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\peppa_pig_face\\facer.py": 1696497788.6561317, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\peppa_pig_face\\face_detector.py": 1696497788.6541314, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\peppa_pig_face\\face_landmark.py": 1696497788.6551151, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\peppa_pig_face\\LK\\lk.py": 1696497788.6571348, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\retinaface\\detection.py": 1696497788.659116, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\retinaface\\models\\net.py": 1696497788.662118, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\retinaface\\models\\retinaface.py": 1696497788.662118, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\retinaface\\utils.py": 1696497788.6601148, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\damofd_detect.py": 1696497788.664131, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\mmdet_patch\\core\\bbox\\transforms.py": 1696497788.66915, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\mmdet_patch\\core\\post_processing\\bbox_nms.py": 1696497788.6711702, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\mmdet_patch\\datasets\\pipelines\\auto_augment.py": 1696497788.6741495, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\mmdet_patch\\datasets\\pipelines\\formating.py": 1696497788.6751683, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\mmdet_patch\\datasets\\pipelines\\loading.py": 1696497788.6751683, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\mmdet_patch\\datasets\\pipelines\\transforms.py": 1696497788.6771507, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\mmdet_patch\\datasets\\retinaface.py": 1696497788.6721745, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\mmdet_patch\\models\\backbones\\master_net.py": 1696497788.6801682, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\mmdet_patch\\models\\backbones\\mobilenet.py": 1696497788.6801682, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\mmdet_patch\\models\\backbones\\resnet.py": 1696497788.6811516, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\mmdet_patch\\models\\dense_heads\\scrfd_head.py": 1696497788.6831515, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\mmdet_patch\\models\\detectors\\base.py": 1696497788.6851497, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\mmdet_patch\\models\\detectors\\scrfd.py": 1696497788.6861672, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\mmdet_patch\\models\\detectors\\single_stage.py": 1696497788.6861672, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\mmdet_patch\\models\\detectors\\tinymog.py": 1696497788.6871495, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\preprocessor.py": 1696497788.664131, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\scrfd_detect.py": 1696497788.6651146, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\scrfd\\tinymog_detect.py": 1696497788.6661146, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\ulfd_slim\\detection.py": 1696497788.6891692, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\ulfd_slim\\vision\\box_utils.py": 1696497788.69017, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\ulfd_slim\\vision\\mb_tiny.py": 1696497788.6911495, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\ulfd_slim\\vision\\ssd\\data_preprocessing.py": 1696497788.6941516, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\ulfd_slim\\vision\\ssd\\fd_config.py": 1696497788.6941516, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\ulfd_slim\\vision\\ssd\\mb_tiny_fd.py": 1696497788.6951494, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\ulfd_slim\\vision\\ssd\\predictor.py": 1696497788.6961498, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\ulfd_slim\\vision\\ssd\\ssd.py": 1696497788.6971686, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_detection\\ulfd_slim\\vision\\transforms.py": 1696497788.6921494, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_emotion\\efficient\\model.py": 1696497788.701167, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_emotion\\efficient\\utils.py": 1696497788.7021701, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_emotion\\emotion_infer.py": 1696497788.698152, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_emotion\\emotion_model.py": 1696497788.69915, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_emotion\\face_alignment\\face.py": 1696497788.703169, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_emotion\\face_alignment\\face_align.py": 1696497788.7041714, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_generation\\op\\conv2d_gradfix.py": 1696497788.707149, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_generation\\op\\fused_act.py": 1696497788.70815, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_generation\\op\\upfirdn2d.py": 1696497788.709151, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_generation\\stylegan2.py": 1696497788.705168, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_human_hand_detection\\det_infer.py": 1696497788.7101765, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_human_hand_detection\\ghost_pan.py": 1696497788.7111504, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_human_hand_detection\\nanodet_plus_head.py": 1696497788.7121665, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_human_hand_detection\\one_stage_detector.py": 1696497788.7121665, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_human_hand_detection\\shufflenetv2.py": 1696497788.7131655, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_human_hand_detection\\utils.py": 1696497788.7141497, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_recognition\\align_face.py": 1696497788.7151647, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_recognition\\torchkit\\backbone\\arcface_backbone.py": 1696497788.7191694, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_recognition\\torchkit\\backbone\\common.py": 1696497788.7191694, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_recognition\\torchkit\\backbone\\facemask_backbone.py": 1696497788.7201505, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_recognition\\torchkit\\backbone\\model_irse.py": 1696497788.7211504, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_recognition\\torchkit\\backbone\\model_resnet.py": 1696497788.7221503, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_recognition\\torchkit\\rts_backbone.py": 1696497788.7171493, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\models\\bfm.py": 1696497788.7251527, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\models\\de_retouching_module.py": 1696497788.7261515, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\models\\facelandmark\\large_base_lmks_infer.py": 1696497788.7331538, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\models\\facelandmark\\nets\\large_base_lmks_net.py": 1696497788.735151, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\models\\facelandmark\\nets\\large_eyeball_net.py": 1696497788.7361522, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\models\\facerecon_model.py": 1696497788.7271519, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\models\\losses.py": 1696497788.7281504, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\models\\networks.py": 1696497788.72915, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\models\\nv_diffrast.py": 1696497788.7301514, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\models\\opt.py": 1696497788.7301514, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\models\\pix2pix\\networks.py": 1696497788.7371516, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\models\\pix2pix\\pix2pix_model.py": 1696497788.7381494, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\models\\pix2pix\\pix2pix_options.py": 1696497788.7391503, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\models\\renderer.py": 1696497788.7311497, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\models\\unet.py": 1696497788.7321503, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\face_reconstruction\\utils.py": 1696497788.72315, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\facial_expression_recognition\\fer\\facial_expression_recognition.py": 1696497788.741153, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\facial_expression_recognition\\fer\\transforms.py": 1696497788.742152, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\facial_expression_recognition\\fer\\vgg.py": 1696497788.7431514, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\facial_landmark_confidence\\flc\\facial_landmark_confidence.py": 1696497788.7451503, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\facial_landmark_confidence\\flc\\manual_landmark_net.py": 1696497788.7461498, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\hand_static\\hand_model.py": 1696497788.7481503, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\hand_static\\networks.py": 1696497788.7481503, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_image_generation\\generators\\base_function.py": 1696497788.7511754, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_image_generation\\generators\\base_module.py": 1696497788.7521503, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_image_generation\\generators\\conv2d_gradfix.py": 1696497788.75315, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_image_generation\\generators\\extraction_distribution_model_flow25.py": 1696497788.75415, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_image_generation\\generators\\flow_module.py": 1696497788.7551703, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_image_generation\\generators\\tps.py": 1696497788.7551703, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_image_generation\\generators\\wavelet_module.py": 1696497788.756151, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_image_generation\\human_image_generation_infer.py": 1696497788.7501507, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_reconstruction\\models\\detectors.py": 1696497788.7631693, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_reconstruction\\models\\Embedding.py": 1696497788.7591505, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_reconstruction\\models\\geometry.py": 1696497788.7631693, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_reconstruction\\models\\human_segmenter.py": 1696497788.7641504, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_reconstruction\\models\\networks.py": 1696497788.7651496, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_reconstruction\\models\\PixToMesh.py": 1696497788.760151, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_reconstruction\\models\\Res_backbone.py": 1696497788.7611675, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_reconstruction\\models\\Surface_head.py": 1696497788.7621682, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_reconstruction\\Reconstruction.py": 1696497788.7571516, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\human_reconstruction\\utils.py": 1696497788.7581499, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_binary_quant_classification\\binary_quant_model.py": 1696497788.7668674, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_binary_quant_classification\\bnext.py": 1696497788.7668674, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_body_reshaping\\image_body_reshaping.py": 1696497788.7688565, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_body_reshaping\\model.py": 1696497788.769877, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_body_reshaping\\person_info.py": 1696497788.769877, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_body_reshaping\\pose_estimator\\body.py": 1696497788.7728715, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_body_reshaping\\pose_estimator\\model.py": 1696497788.7728715, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_body_reshaping\\pose_estimator\\util.py": 1696497788.7738824, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_body_reshaping\\slim_utils.py": 1696497788.7708614, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_classification\\backbones\\beit_v2.py": 1696497788.7788763, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_classification\\backbones\\nextvit.py": 1696497788.7798567, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_classification\\mmcls_model.py": 1696497788.7758567, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_classification\\resnet50_cc.py": 1696497788.7768598, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_classification\\utils.py": 1696497788.7778792, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_colorization\\ddcolor\\ddcolor.py": 1696497788.7898567, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_colorization\\ddcolor\\ddcolor_for_image_colorization.py": 1696497788.7908561, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_colorization\\ddcolor\\loss.py": 1696497788.7908561, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_colorization\\ddcolor\\utils\\convnext.py": 1696497788.7928565, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_colorization\\ddcolor\\utils\\position_encoding.py": 1696497788.7928565, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_colorization\\ddcolor\\utils\\transformer_utils.py": 1696497788.7938564, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_colorization\\ddcolor\\utils\\unet.py": 1696497788.7948563, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_colorization\\ddcolor\\utils\\vgg.py": 1696497788.795856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_colorization\\unet\\unet.py": 1696497788.796878, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_colorization\\unet\\utils.py": 1696497788.7978573, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_color_enhance\\adaint\\adaint.py": 1696497788.7838593, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_color_enhance\\csrnet.py": 1696497788.7818606, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_color_enhance\\deeplpf\\deeplpfnet.py": 1696497788.7868576, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_color_enhance\\deeplpf\\deeplpf_image_color_enhance.py": 1696497788.785857, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_color_enhance\\image_color_enhance.py": 1696497788.7818606, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_debanding\\rrdb\\rrdb_image_debanding.py": 1696497788.799856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_deblur\\nafnet_for_image_deblur.py": 1696497788.8018558, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\defrcn_for_fewshot.py": 1696497788.8028562, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\evaluation\\coco_evaluation.py": 1696497788.804856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\evaluation\\evaluator.py": 1696497788.805856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\evaluation\\pascal_voc_evaluation.py": 1696497788.8068569, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\models\\calibration_layer.py": 1696497788.8088572, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\models\\defrcn.py": 1696497788.8088572, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\models\\fast_rcnn.py": 1696497788.8098595, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\models\\gdl.py": 1696497788.8108594, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\models\\resnet.py": 1696497788.811857, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\models\\roi_heads.py": 1696497788.811857, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\utils\\coco_register.py": 1696497788.813856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\utils\\configuration_mapper.py": 1696497788.814856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\utils\\model_surgery_op.py": 1696497788.814856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\utils\\register_data.py": 1696497788.8158562, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\utils\\requirements_check.py": 1696497788.8168564, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_defrcn_fewshot\\utils\\voc_register.py": 1696497788.817856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_denoise\\nafnet\\arch_util.py": 1696497788.820855, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_denoise\\nafnet\\NAFNet_arch.py": 1696497788.8198562, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_denoise\\nafnet_for_image_denoise.py": 1696497788.8188562, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_depth_estimation\\networks\\newcrf_depth.py": 1696497788.8238547, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_depth_estimation\\networks\\newcrf_layers.py": 1696497788.8248546, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_depth_estimation\\networks\\newcrf_utils.py": 1696497788.8258553, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_depth_estimation\\networks\\swin_transformer.py": 1696497788.8268554, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_depth_estimation\\networks\\uper_crf_head.py": 1696497788.8278584, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_depth_estimation\\newcrfs_model.py": 1696497788.8228548, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_depth_estimation_bts\\depth_estimation_bts_model.py": 1696497788.8288577, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_depth_estimation_bts\\networks\\bts_model.py": 1696497788.8308566, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_depth_estimation_bts\\networks\\decoder.py": 1696497788.8318558, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_depth_estimation_bts\\networks\\encoder.py": 1696497788.8318558, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_depth_estimation_bts\\networks\\utils.py": 1696497788.8328564, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_driving_perception\\image_driving_percetion_model.py": 1696497788.8348567, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_driving_perception\\preprocessor.py": 1696497788.8358567, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_driving_perception\\utils.py": 1696497788.8368566, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_editing\\masactrl.py": 1696497788.8388584, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_editing\\masactrl_utils.py": 1696497788.8388584, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_face_fusion\\facegan\\face_gan.py": 1696497788.8418572, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_face_fusion\\facegan\\gpen_model.py": 1696497788.842857, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_face_fusion\\facegan\\op\\conv2d_gradfix.py": 1696497788.8448558, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_face_fusion\\facegan\\op\\fused_act.py": 1696497788.8448558, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_face_fusion\\facegan\\op\\upfirdn2d.py": 1696497788.845856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_face_fusion\\facelib\\align_trans.py": 1696497788.8478572, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_face_fusion\\facelib\\matlab_cp2tform.py": 1696497788.8488593, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_face_fusion\\image_face_fusion.py": 1696497788.8408558, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_face_fusion\\network\\aad_layer.py": 1696497788.8508575, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_face_fusion\\network\\aei_flow_net.py": 1696497788.851857, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_face_fusion\\network\\bfm.py": 1696497788.852856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_face_fusion\\network\\dense_motion.py": 1696497788.8538556, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_face_fusion\\network\\facerecon_model.py": 1696497788.854859, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_face_fusion\\network\\model_irse.py": 1696497788.854859, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_face_fusion\\network\\ops.py": 1696497788.8558607, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_human_parsing\\backbone\\deeplab_resnet.py": 1696497788.860857, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_human_parsing\\m2fp\\m2fp_decoder.py": 1696497788.8618562, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_human_parsing\\m2fp\\m2fp_encoder.py": 1696497788.8628564, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_human_parsing\\m2fp_net.py": 1696497788.8578587, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_human_parsing\\parsing_utils.py": 1696497788.8588564, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_inpainting\\base.py": 1696497788.8638568, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_inpainting\\default.py": 1696497788.864856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_inpainting\\model.py": 1696497788.865856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_inpainting\\modules\\ade20k\\base.py": 1696497788.8738558, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_inpainting\\modules\\ade20k\\resnet.py": 1696497788.8748565, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_inpainting\\modules\\adversarial.py": 1696497788.8678591, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_inpainting\\modules\\feature_matching.py": 1696497788.868857, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_inpainting\\modules\\ffc.py": 1696497788.8698568, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_inpainting\\modules\\inception.py": 1696497788.8708591, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_inpainting\\modules\\perceptual.py": 1696497788.8718567, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_inpainting\\modules\\pix2pixhd.py": 1696497788.8718567, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_inpainting\\refinement.py": 1696497788.8668559, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\backbones\\resnet.py": 1696497788.8818564, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\backbones\\swin_transformer.py": 1696497788.882856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\cascade_mask_rcnn_swin.py": 1696497788.8758583, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\datasets\\transforms.py": 1696497788.8848562, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\fastinst\\fastinst_decoder.py": 1696497788.8858564, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\fastinst\\fastinst_encoder.py": 1696497788.8868566, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\fastinst_model.py": 1696497788.8768568, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\maskdino\\dino_decoder.py": 1696497788.888857, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\maskdino\\maskdino_decoder.py": 1696497788.8898592, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\maskdino\\maskdino_encoder.py": 1696497788.8908803, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\maskdino\\ms_deform_attn.py": 1696497788.8918772, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\maskdino\\position_encoding.py": 1696497788.8918772, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\maskdino\\utils.py": 1696497788.8928573, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\maskdino_model.py": 1696497788.8778563, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\maskdino_swin.py": 1696497788.878857, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\model.py": 1696497788.8798587, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_instance_segmentation\\postprocess_utils.py": 1696497788.880857, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_matching\\config\\default.py": 1696497788.895856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_matching\\loftr_quadtree\\backbone\\resnet_fpn.py": 1696497788.899858, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_matching\\loftr_quadtree\\loftr.py": 1696497788.8978605, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_matching\\loftr_quadtree\\loftr_module\\fine_preprocess.py": 1696497788.900876, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_matching\\loftr_quadtree\\loftr_module\\linear_attention.py": 1696497788.9018567, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_matching\\loftr_quadtree\\loftr_module\\quadtree_attention.py": 1696497788.902856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_matching\\loftr_quadtree\\loftr_module\\transformer.py": 1696497788.9039097, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_matching\\loftr_quadtree\\utils\\coarse_matching.py": 1696497788.9054205, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_matching\\loftr_quadtree\\utils\\fine_matching.py": 1696497788.9054205, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_matching\\loftr_quadtree\\utils\\position_encoding.py": 1696497788.9064207, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_matching\\quadtree_attention_model.py": 1696497788.8948562, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_matching\\utils\\misc.py": 1696497788.9084218, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_mvs_depth_estimation\\casmvs_model.py": 1696497788.9104207, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_mvs_depth_estimation\\cas_mvsnet.py": 1696497788.9094217, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_mvs_depth_estimation\\colmap2mvsnet.py": 1696497788.9114242, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_mvs_depth_estimation\\depth_filter.py": 1696497788.9124234, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_mvs_depth_estimation\\general_eval_dataset.py": 1696497788.9134243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_mvs_depth_estimation\\module.py": 1696497788.914423, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_mvs_depth_estimation\\utils.py": 1696497788.914423, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_paintbyexample\\model.py": 1696497788.9164236, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_panoptic_segmentation\\panseg_model.py": 1696497788.9174237, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_portrait_enhancement\\align_faces.py": 1696497788.9194233, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_portrait_enhancement\\eqface\\fqa.py": 1696497788.9234233, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_portrait_enhancement\\eqface\\model_resnet.py": 1696497788.9234233, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_portrait_enhancement\\gpen.py": 1696497788.9204218, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_portrait_enhancement\\image_portrait_enhancement.py": 1696497788.9214213, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_portrait_enhancement\\losses\\helpers.py": 1696497788.9254231, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_portrait_enhancement\\losses\\losses.py": 1696497788.9264238, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_portrait_enhancement\\losses\\model_irse.py": 1696497788.9274237, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_portrait_enhancement\\retinaface\\detection.py": 1696497788.928425, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_portrait_enhancement\\retinaface\\models\\net.py": 1696497788.930425, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_portrait_enhancement\\retinaface\\models\\retinaface.py": 1696497788.9314246, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_portrait_enhancement\\retinaface\\utils.py": 1696497788.929424, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_probing_model\\backbone.py": 1696497788.9334245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_probing_model\\model.py": 1696497788.9334245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_probing_model\\utils.py": 1696497788.9344244, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_quality_assessment_degradation\\degradation_model.py": 1696497788.9364245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_quality_assessment_degradation\\image_quality_assessment_degradation.py": 1696497788.9374247, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_quality_assessment_man\\image_quality_assessment_man.py": 1696497788.9394248, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_quality_assessment_man\\maniqa.py": 1696497788.9394248, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_quality_assessment_man\\swin.py": 1696497788.9404233, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_quality_assessment_mos\\backbones\\resnet.py": 1696497788.946423, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_quality_assessment_mos\\censeo_ivqa_model.py": 1696497788.9424238, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_quality_assessment_mos\\heads\\simple_head.py": 1696497788.9474218, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_quality_assessment_mos\\image_quality_assessment_mos.py": 1696497788.943424, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_reid_person\\pass_model.py": 1696497788.949422, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_reid_person\\transreid_model.py": 1696497788.9504218, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_restoration\\demoire_models\\nets.py": 1696497788.953421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_restoration\\image_restoration_model.py": 1696497788.9514215, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\ddpm_seg\\data_util.py": 1696497788.957435, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\ddpm_seg\\feature_extractors.py": 1696497788.958423, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\ddpm_seg\\pixel_classifier.py": 1696497788.959422, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\ddpm_seg\\utils.py": 1696497788.961424, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\ddpm_segmentation_model.py": 1696497788.955423, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\pan_merge\\base_panoptic_fusion_head.py": 1696497788.9624255, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\pan_merge\\maskformer_semantic_head.py": 1696497788.963425, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\semantic_seg_model.py": 1696497788.955423, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\vit_adapter\\models\\backbone\\adapter_modules.py": 1696497788.9664392, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\vit_adapter\\models\\backbone\\base\\beit.py": 1696497788.969437, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\vit_adapter\\models\\backbone\\beit_adapter.py": 1696497788.9674406, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\vit_adapter\\models\\decode_heads\\base_decode_head.py": 1696497788.9714406, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\vit_adapter\\models\\decode_heads\\mask2former_head_from_mmseg.py": 1696497788.9714406, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\vit_adapter\\models\\segmentors\\base_segmentor.py": 1696497788.9734206, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\vit_adapter\\models\\segmentors\\encoder_decoder_mask2former.py": 1696497788.9744205, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\vit_adapter\\utils\\builder.py": 1696497788.9764228, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\vit_adapter\\utils\\data_process_func.py": 1696497788.977426, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_semantic_segmentation\\vit_adapter\\utils\\seg_func.py": 1696497788.977426, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_skychange\\preprocessor.py": 1696497788.9794397, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_skychange\\ptsemseg\\BlockModules.py": 1696497788.9814205, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_skychange\\ptsemseg\\hrnet_backnone.py": 1696497788.9834385, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_skychange\\ptsemseg\\hrnet_super_and_ocr.py": 1696497788.9834385, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_skychange\\ptsemseg\\unet.py": 1696497788.98444, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_skychange\\skychange.py": 1696497788.9794397, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_skychange\\skychange_model.py": 1696497788.9804204, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_super_resolution_pasd\\attention.py": 1696497788.9864366, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_super_resolution_pasd\\controlnet.py": 1696497788.9874487, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_super_resolution_pasd\\misc.py": 1696497788.9874487, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_super_resolution_pasd\\transformer_2d.py": 1696497788.9884217, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_super_resolution_pasd\\unet_2d_blocks.py": 1696497789.0074227, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_super_resolution_pasd\\unet_2d_condition.py": 1696497789.0094228, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_generation\\data\\transforms.py": 1696497789.014423, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_generation\\model.py": 1696497789.0114298, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_generation\\models\\autoencoder.py": 1696497789.0154228, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_generation\\models\\clip.py": 1696497789.017423, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_generation\\ops\\diffusion.py": 1696497789.0184233, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_generation\\ops\\losses.py": 1696497789.0194218, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_translation\\data\\transforms.py": 1696497789.0224202, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_translation\\models\\autoencoder.py": 1696497789.02444, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_translation\\models\\clip.py": 1696497789.02444, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_translation\\model_translation.py": 1696497789.021421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_translation\\ops\\apps.py": 1696497789.0264225, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_translation\\ops\\degradation.py": 1696497789.0274441, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_translation\\ops\\diffusion.py": 1696497789.0284224, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_translation\\ops\\losses.py": 1696497789.0294409, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_translation\\ops\\metrics.py": 1696497789.0304422, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_translation\\ops\\random_color.py": 1696497789.0304422, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_translation\\ops\\random_mask.py": 1696497789.0314212, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_translation\\ops\\svd.py": 1696497789.0324402, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_to_image_translation\\ops\\utils.py": 1696497789.0324402, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_try_on\\generator.py": 1696497789.034421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_try_on\\landmark.py": 1696497789.0354373, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_try_on\\try_on_infer.py": 1696497789.0364244, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\image_try_on\\warping.py": 1696497789.0374408, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\indoor_layout_estimation\\networks\\backbone\\resnet_DA.py": 1696497789.0424428, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\indoor_layout_estimation\\networks\\backbone\\vit_horizon_pry_image.py": 1696497789.0434434, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\indoor_layout_estimation\\networks\\misc\\fourier.py": 1696497789.045438, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\indoor_layout_estimation\\networks\\misc\\panostretch.py": 1696497789.0464401, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\indoor_layout_estimation\\networks\\misc\\post_proc.py": 1696497789.0464401, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\indoor_layout_estimation\\networks\\modality\\layout.py": 1696497789.0484207, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\indoor_layout_estimation\\networks\\panovit.py": 1696497789.040423, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\indoor_layout_estimation\\networks\\utils.py": 1696497789.0414371, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\indoor_layout_estimation\\panovit.py": 1696497789.0384257, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\language_guided_video_summarization\\summarizer.py": 1696497789.0504375, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\language_guided_video_summarization\\transformer\\layers.py": 1696497789.0514207, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\language_guided_video_summarization\\transformer\\models.py": 1696497789.0524206, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\language_guided_video_summarization\\transformer\\modules.py": 1696497789.0534372, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\language_guided_video_summarization\\transformer\\sub_layers.py": 1696497789.0534372, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\motion_generation\\model.py": 1696497789.0554204, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\motion_generation\\modules\\cfg_sampler.py": 1696497789.0574226, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\motion_generation\\modules\\gaussian_diffusion.py": 1696497789.0574226, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\motion_generation\\modules\\mdm.py": 1696497789.058421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\motion_generation\\modules\\respace.py": 1696497789.0594416, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\motion_generation\\modules\\rotation2xyz.py": 1696497789.060438, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\motion_generation\\modules\\smpl.py": 1696497789.060438, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\movie_scene_segmentation\\get_model.py": 1696497789.062421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\movie_scene_segmentation\\model.py": 1696497789.063421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\movie_scene_segmentation\\utils\\head.py": 1696497789.06444, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\movie_scene_segmentation\\utils\\save_op.py": 1696497789.0654209, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\movie_scene_segmentation\\utils\\shot_encoder.py": 1696497789.0664413, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\movie_scene_segmentation\\utils\\trn.py": 1696497789.0674245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_4k\\dataloader\\load_blender.py": 1696497789.0704432, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_4k\\dataloader\\load_data.py": 1696497789.0714211, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_4k\\dataloader\\load_llff.py": 1696497789.0724373, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_4k\\dataloader\\load_tankstemple.py": 1696497789.073439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_4k\\dataloader\\read_write_model.py": 1696497789.073439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_4k\\nerf_preprocess.py": 1696497789.0684204, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_4k\\nerf_recon_4k.py": 1696497789.0694425, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_4k\\network\\dvgo.py": 1696497789.0754359, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_4k\\network\\utils.py": 1696497789.0764222, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_acc\\dataloader\\nerf_dataset.py": 1696497789.0804365, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_acc\\dataloader\\read_write_model.py": 1696497789.081439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_acc\\nerf_preprocess.py": 1696497789.0784416, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_acc\\nerf_recon_acc.py": 1696497789.0784416, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_acc\\network\\nerf.py": 1696497789.0824401, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_acc\\network\\segmenter.py": 1696497789.0834224, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_acc\\network\\utils.py": 1696497789.0844376, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_vq_compression\\dataloader\\blender.py": 1696497789.0884416, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_vq_compression\\dataloader\\llff.py": 1696497789.089421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_vq_compression\\dataloader\\nsvf.py": 1696497789.090438, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_vq_compression\\dataloader\\ray_utils.py": 1696497789.0914378, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_vq_compression\\dataloader\\tankstemple.py": 1696497789.0914378, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_vq_compression\\nerf_recon_vq_compression.py": 1696497789.0854242, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_vq_compression\\network\\tensorBase.py": 1696497789.094442, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_vq_compression\\network\\tensoRF.py": 1696497789.093421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_vq_compression\\network\\tensoRF_VQ.py": 1696497789.094442, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_vq_compression\\network\\weighted_vq.py": 1696497789.0954468, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_vq_compression\\renderer.py": 1696497789.086442, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\nerf_recon_vq_compression\\utils.py": 1696497789.0874412, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection\\mmdet_model.py": 1696497789.097421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection\\mmdet_ms\\backbones\\vit.py": 1696497789.0994406, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection\\mmdet_ms\\dense_heads\\anchor_head.py": 1696497789.1014395, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection\\mmdet_ms\\dense_heads\\rpn_head.py": 1696497789.102425, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection\\mmdet_ms\\necks\\fpn.py": 1696497789.1034212, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection\\mmdet_ms\\roi_heads\\bbox_heads\\convfc_bbox_head.py": 1696497789.1064403, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection\\mmdet_ms\\roi_heads\\mask_heads\\fcn_mask_head.py": 1696497789.107421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection\\mmdet_ms\\utils\\checkpoint.py": 1696497789.1094224, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection\\mmdet_ms\\utils\\convModule_norm.py": 1696497789.1104403, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\depe_detect.py": 1696497789.112421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\mmdet3d_plugin\\core\\bbox\\assigners\\hungarian_assigner_3d.py": 1696497789.1174233, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\mmdet3d_plugin\\core\\bbox\\coders\\nms_free_coder.py": 1696497789.1194212, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\mmdet3d_plugin\\core\\bbox\\match_costs\\match_cost.py": 1696497789.1214218, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\mmdet3d_plugin\\core\\bbox\\util.py": 1696497789.1164365, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\mmdet3d_plugin\\datasets\\nuscenes_dataset.py": 1696497789.1234405, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\mmdet3d_plugin\\datasets\\pipelines\\loading.py": 1696497789.1244218, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\mmdet3d_plugin\\datasets\\pipelines\\transform_3d.py": 1696497789.1254213, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\mmdet3d_plugin\\models\\backbones\\vovnet.py": 1696497789.127437, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\mmdet3d_plugin\\models\\dense_heads\\depth_net.py": 1696497789.1294205, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\mmdet3d_plugin\\models\\dense_heads\\petrv2_dednhead.py": 1696497789.1304226, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\mmdet3d_plugin\\models\\detectors\\petr3d.py": 1696497789.1324413, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\mmdet3d_plugin\\models\\necks\\cp_fpn.py": 1696497789.1334398, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\mmdet3d_plugin\\models\\utils\\petr_transformer.py": 1696497789.1354208, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\mmdet3d_plugin\\models\\utils\\positional_encoding.py": 1696497789.136421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\object_detection_3d\\depe\\result_vis.py": 1696497789.1134405, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_detection\\model.py": 1696497789.1374211, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_detection\\modules\\dbnet.py": 1696497789.1414206, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_detection\\modules\\layers.py": 1696497789.1414206, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_detection\\modules\\mix_ops.py": 1696497789.1424203, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_detection\\modules\\proxyless.py": 1696497789.143421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_detection\\modules\\seg_detector_loss.py": 1696497789.1444206, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_detection\\preprocessor.py": 1696497789.1384208, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_detection\\utils.py": 1696497789.139421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_recognition\\model.py": 1696497789.1464205, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_recognition\\modules\\ConvNextViT\\convnext.py": 1696497789.151421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_recognition\\modules\\ConvNextViT\\main_model.py": 1696497789.151421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_recognition\\modules\\ConvNextViT\\timm_tinyc.py": 1696497789.152421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_recognition\\modules\\ConvNextViT\\vitstr.py": 1696497789.1534207, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_recognition\\modules\\CRNN\\main_model.py": 1696497789.1494207, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_recognition\\modules\\LightweightEdge\\main_model.py": 1696497789.1554205, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_recognition\\modules\\LightweightEdge\\nas_block\\layers.py": 1696497789.1574202, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_recognition\\modules\\LightweightEdge\\nas_block\\mix_ops.py": 1696497789.1574202, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_recognition\\modules\\LightweightEdge\\nas_block\\proxyless.py": 1696497789.1584206, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\ocr_recognition\\preprocessor.py": 1696497789.1474204, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\open_vocabulary_detection_vild\\vild.py": 1696497789.1604204, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\panorama_depth_estimation\\networks\\equi.py": 1696497789.164425, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\panorama_depth_estimation\\networks\\layers.py": 1696497789.164425, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\panorama_depth_estimation\\networks\\mobilenet.py": 1696497789.1654243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\panorama_depth_estimation\\networks\\resnet.py": 1696497789.1664243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\panorama_depth_estimation\\networks\\unifuse.py": 1696497789.1675034, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\panorama_depth_estimation\\networks\\util.py": 1696497789.1685004, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\panorama_depth_estimation\\unifuse_model.py": 1696497789.1624253, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\pedestrian_attribute_recognition\\model.py": 1696497789.169493, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\pointcloud_sceneflow_estimation\\common.py": 1696497789.1714954, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\pointcloud_sceneflow_estimation\\pointnet2_utils.py": 1696497789.1714954, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\pointcloud_sceneflow_estimation\\rcp_model.py": 1696497789.1724331, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\pointcloud_sceneflow_estimation\\sf_rcp.py": 1696497789.1734233, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\product_retrieval_embedding\\item_detection.py": 1696497789.175502, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\product_retrieval_embedding\\item_embedding.py": 1696497789.175502, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\product_retrieval_embedding\\item_model.py": 1696497789.1764226, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\product_segmentation\\net.py": 1696497789.1784537, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\product_segmentation\\seg_infer.py": 1696497789.1795013, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\referring_video_object_segmentation\\model.py": 1696497789.1805227, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\referring_video_object_segmentation\\utils\\backbone.py": 1696497789.182421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\referring_video_object_segmentation\\utils\\criterion.py": 1696497789.183496, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\referring_video_object_segmentation\\utils\\matcher.py": 1696497789.1844935, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\referring_video_object_segmentation\\utils\\misc.py": 1696497789.1844935, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\referring_video_object_segmentation\\utils\\mttr.py": 1696497789.185423, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\referring_video_object_segmentation\\utils\\multimodal_transformer.py": 1696497789.1864514, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\referring_video_object_segmentation\\utils\\position_encoding_2d.py": 1696497789.1864514, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\referring_video_object_segmentation\\utils\\postprocessing.py": 1696497789.1875, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\referring_video_object_segmentation\\utils\\segmentation.py": 1696497789.18847, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\referring_video_object_segmentation\\utils\\swin_transformer.py": 1696497789.189508, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\robust_image_classification\\easyrobust_model.py": 1696497789.190437, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\s2net_panorama_depth_estimation\\networks\\config.py": 1696497789.1934237, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\s2net_panorama_depth_estimation\\networks\\decoder.py": 1696497789.1944416, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\s2net_panorama_depth_estimation\\networks\\model.py": 1696497789.1954212, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\s2net_panorama_depth_estimation\\networks\\resnet.py": 1696497789.196441, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\s2net_panorama_depth_estimation\\networks\\swin_transformer.py": 1696497789.19744, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\s2net_panorama_depth_estimation\\networks\\util_helper.py": 1696497789.19744, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\s2net_panorama_depth_estimation\\s2net_model.py": 1696497789.1924214, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\salient_detection\\models\\backbone\\Res2Net_v1b.py": 1696497789.203441, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\salient_detection\\models\\modules.py": 1696497789.2004397, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\salient_detection\\models\\senet.py": 1696497789.2014406, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\salient_detection\\models\\u2net.py": 1696497789.2024212, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\salient_detection\\models\\utils.py": 1696497789.2024212, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\salient_detection\\salient_model.py": 1696497789.199421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\shop_segmentation\\common.py": 1696497789.206439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\shop_segmentation\\head_fpn.py": 1696497789.206439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\shop_segmentation\\models.py": 1696497789.207421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\shop_segmentation\\neck_fpn.py": 1696497789.2084227, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\shop_segmentation\\shop_seg_base.py": 1696497789.209445, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\shop_segmentation\\shop_seg_model.py": 1696497789.2104404, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\shop_segmentation\\utils.py": 1696497789.2104404, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\skin_retouching\\detection_model\\detection_module.py": 1696497789.2154374, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\skin_retouching\\detection_model\\detection_unet_in.py": 1696497789.2154374, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\skin_retouching\\inpainting_model\\gconv.py": 1696497789.217439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\skin_retouching\\inpainting_model\\inpainting_unet.py": 1696497789.218422, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\skin_retouching\\retinaface\\box_utils.py": 1696497789.219442, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\skin_retouching\\retinaface\\net.py": 1696497789.2204213, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\skin_retouching\\retinaface\\network.py": 1696497789.2204213, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\skin_retouching\\retinaface\\predict_single.py": 1696497789.2214365, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\skin_retouching\\retinaface\\prior_box.py": 1696497789.2224212, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\skin_retouching\\retinaface\\utils.py": 1696497789.223439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\skin_retouching\\unet_deploy.py": 1696497789.2124379, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\skin_retouching\\utils.py": 1696497789.2124379, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\skin_retouching\\weights_init.py": 1696497789.2134404, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\stream_yolo\\data\\data_augment.py": 1696497789.2264216, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\stream_yolo\\exp\\base_exp.py": 1696497789.2284229, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\stream_yolo\\exp\\build.py": 1696497789.2284229, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\stream_yolo\\exp\\default\\streamyolo.py": 1696497789.2314205, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\stream_yolo\\exp\\yolox_base.py": 1696497789.2294211, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\stream_yolo\\models\\darknet.py": 1696497789.2324235, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\stream_yolo\\models\\dfp_pafpn.py": 1696497789.2334232, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\stream_yolo\\models\\network_blocks.py": 1696497789.234422, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\stream_yolo\\models\\streamyolo.py": 1696497789.2354434, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\stream_yolo\\models\\tal_head.py": 1696497789.2354434, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\stream_yolo\\realtime_video_detector.py": 1696497789.2244444, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\stream_yolo\\utils\\boxes.py": 1696497789.2374203, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\stream_yolo\\utils\\format.py": 1696497789.2384422, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\super_resolution\\arch_util.py": 1696497789.2394397, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\super_resolution\\ecb.py": 1696497789.2404213, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\super_resolution\\ecbsr_model.py": 1696497789.2414393, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\super_resolution\\rrdbnet_arch.py": 1696497789.2424219, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\surface_recon_common\\dataset.py": 1696497789.2434216, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\surface_recon_common\\fields.py": 1696497789.244439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\surface_recon_common\\renderer.py": 1696497789.2454417, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\surface_recon_common\\surface_recon_common.py": 1696497789.2454417, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\surface_recon_common\\utils.py": 1696497789.246421, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\table_recognition\\lineless_table_process.py": 1696497789.2484226, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\table_recognition\\model_lore.py": 1696497789.2494438, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\table_recognition\\modules\\lore_detector.py": 1696497789.2504213, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\table_recognition\\modules\\lore_processor.py": 1696497789.2514207, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\text_driven_segmentation\\clip.py": 1696497789.2524204, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\text_driven_segmentation\\lseg_base.py": 1696497789.2534206, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\text_driven_segmentation\\lseg_blocks.py": 1696497789.2544212, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\text_driven_segmentation\\lseg_model.py": 1696497789.255423, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\text_driven_segmentation\\lseg_net.py": 1696497789.255423, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\text_driven_segmentation\\lseg_vit.py": 1696497789.2574244, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\text_driven_segmentation\\model.py": 1696497789.2584229, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\text_driven_segmentation\\simple_tokenizer.py": 1696497789.2584229, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\text_to_360panorama_image\\pipeline_base.py": 1696497789.260424, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\text_to_360panorama_image\\pipeline_sr.py": 1696497789.2614224, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_classfication\\basic_blocks.py": 1696497789.2634456, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_classfication\\global_utils.py": 1696497789.2644415, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_classfication\\master_net.py": 1696497789.2644415, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_classfication\\model_zoo.py": 1696497789.2654212, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_classfication\\plain_net_utils.py": 1696497789.2664216, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_classfication\\super_blocks.py": 1696497789.267422, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_classfication\\super_res_idwexkx.py": 1696497789.267422, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_classfication\\super_res_k1kxk1.py": 1696497789.26842, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_classfication\\super_res_kxkx.py": 1696497789.2694254, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\apis\\detector_evaluater.py": 1696497789.2764244, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\apis\\detector_inference.py": 1696497789.2774222, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\augmentations\\box_level_augs\\box_level_augs.py": 1696497789.2804408, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\augmentations\\box_level_augs\\color_augs.py": 1696497789.2814212, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\augmentations\\box_level_augs\\gaussian_maps.py": 1696497789.2824392, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\augmentations\\box_level_augs\\geometric_augs.py": 1696497789.2824392, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\augmentations\\scale_aware_aug.py": 1696497789.279442, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\backbones\\darknet.py": 1696497789.2854397, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\backbones\\tinynas_csp.py": 1696497789.2864237, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\backbones\\tinynas_res.py": 1696497789.2864237, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\core\\base_ops.py": 1696497789.2884219, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\core\\neck_ops.py": 1696497789.2894373, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\core\\ops.py": 1696497789.2904403, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\core\\ota_assigner.py": 1696497789.2904403, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\core\\repvgg_block.py": 1696497789.2914255, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\core\\utils.py": 1696497789.2924216, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\core\\weight_init.py": 1696497789.2934213, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\heads\\gfocal_v2_tiny.py": 1696497789.2944386, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\heads\\zero_head.py": 1696497789.2954211, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\losses\\distill_loss.py": 1696497789.2974257, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\losses\\gfocal_loss.py": 1696497789.2974257, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\necks\\giraffe_config.py": 1696497789.2994227, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\necks\\giraffe_fpn.py": 1696497789.3004208, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\base_models\\necks\\giraffe_fpn_btn.py": 1696497789.301439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\detectors\\detector.py": 1696497789.3034406, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\structures\\bounding_box.py": 1696497789.3044214, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\structures\\boxlist_ops.py": 1696497789.305441, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\structures\\image_list.py": 1696497789.305441, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\utils\\boxes.py": 1696497789.3074458, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\utils\\model_utils.py": 1696497789.3084226, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\damo\\utils\\scheduler.py": 1696497789.3094418, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\detector.py": 1696497789.2714205, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\tinynas_damoyolo.py": 1696497789.2724245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\tinynas_detector.py": 1696497789.2734244, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\tinynas_detection\\utils.py": 1696497789.274443, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_deinterlace\\deinterlace_arch.py": 1696497789.3115468, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_deinterlace\\models\\archs.py": 1696497789.3135438, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_deinterlace\\models\\deep_fourier_upsampling.py": 1696497789.314541, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_deinterlace\\models\\enh.py": 1696497789.314541, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_deinterlace\\models\\fre.py": 1696497789.315524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_deinterlace\\models\\utils.py": 1696497789.3165426, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_deinterlace\\UNet_for_video_deinterlace.py": 1696497789.3104794, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\configs\\default_config.py": 1696497789.3195405, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\dro_model.py": 1696497789.3175282, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\geometry\\camera.py": 1696497789.3215444, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\geometry\\camera_utils.py": 1696497789.3215444, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\geometry\\pose.py": 1696497789.3225253, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\geometry\\pose_utils.py": 1696497789.3235424, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\models\\model_checkpoint.py": 1696497789.3245437, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\models\\model_utils.py": 1696497789.3255253, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\models\\model_wrapper.py": 1696497789.3265252, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\models\\sfm_model_mf.py": 1696497789.327524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\models\\sup_model_mf.py": 1696497789.327524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\networks\\depth_pose\\depth_pose_net.py": 1696497789.3305423, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\networks\\layers\\resnet\\depth_decoder.py": 1696497789.332548, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\networks\\layers\\resnet\\layers.py": 1696497789.3335268, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\networks\\layers\\resnet\\pose_decoder.py": 1696497789.3345253, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\networks\\layers\\resnet\\resnet_encoder.py": 1696497789.3355243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\networks\\optim\\extractor.py": 1696497789.3365245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\networks\\optim\\update.py": 1696497789.3375242, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\utils\\augmentations.py": 1696497789.339524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\utils\\config.py": 1696497789.339524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\utils\\depth.py": 1696497789.3405242, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\utils\\horovod.py": 1696497789.341524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\utils\\image.py": 1696497789.3425238, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\utils\\image_gt.py": 1696497789.3435235, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\utils\\load.py": 1696497789.3445275, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\utils\\misc.py": 1696497789.3445275, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_depth_estimation\\utils\\types.py": 1696497789.3455248, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_frame_interpolation\\flow_model\\corr.py": 1696497789.3505263, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_frame_interpolation\\flow_model\\extractor.py": 1696497789.3515272, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_frame_interpolation\\flow_model\\raft.py": 1696497789.3515272, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_frame_interpolation\\flow_model\\update.py": 1696497789.3535268, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_frame_interpolation\\interp_model\\flow_reversal.py": 1696497789.356529, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_frame_interpolation\\interp_model\\IFNet_swin.py": 1696497789.3545291, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_frame_interpolation\\interp_model\\refinenet_arch.py": 1696497789.3575253, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_frame_interpolation\\interp_model\\transformer_layers.py": 1696497789.3595262, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_frame_interpolation\\interp_model\\UNet.py": 1696497789.3555298, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_frame_interpolation\\utils\\scene_change_detection.py": 1696497789.3615265, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_frame_interpolation\\utils\\utils.py": 1696497789.3615265, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_frame_interpolation\\VFINet_arch.py": 1696497789.3465452, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_frame_interpolation\\VFINet_for_video_frame_interpolation.py": 1696497789.347543, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_human_matting\\model.py": 1696497789.3635285, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_human_matting\\models\\decoder.py": 1696497789.3655272, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_human_matting\\models\\deep_guided_filter.py": 1696497789.3665276, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_human_matting\\models\\effv2.py": 1696497789.367544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_human_matting\\models\\lraspp.py": 1696497789.368546, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_human_matting\\models\\matting.py": 1696497789.3695285, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_inpainting\\inpainting.py": 1696497789.3715258, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_inpainting\\inpainting_model.py": 1696497789.3715258, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_instance_segmentation\\head\\kernel_frame_iter_head.py": 1696497789.375524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_instance_segmentation\\head\\kernel_head.py": 1696497789.3765256, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_instance_segmentation\\head\\kernel_iter_head.py": 1696497789.377524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_instance_segmentation\\head\\kernel_update_head.py": 1696497789.3785431, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_instance_segmentation\\head\\kernel_updator.py": 1696497789.3795276, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_instance_segmentation\\neck\\msdeformattn_decoder.py": 1696497789.380544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_instance_segmentation\\track\\kernel_update_head.py": 1696497789.38254, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_instance_segmentation\\track\\mask_hungarian_assigner.py": 1696497789.3835435, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_instance_segmentation\\utils.py": 1696497789.3735406, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_instance_segmentation\\video_knet.py": 1696497789.374544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_multi_object_tracking\\models\\common.py": 1696497789.3855422, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_multi_object_tracking\\models\\decode.py": 1696497789.386545, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_multi_object_tracking\\models\\model.py": 1696497789.386545, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_multi_object_tracking\\models\\yolo.py": 1696497789.3875241, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_multi_object_tracking\\tracker\\basetrack.py": 1696497789.3885427, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_multi_object_tracking\\tracker\\matching.py": 1696497789.38954, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_multi_object_tracking\\tracker\\multitracker.py": 1696497789.3905413, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_multi_object_tracking\\utils\\image.py": 1696497789.3915396, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_multi_object_tracking\\utils\\kalman_filter.py": 1696497789.392526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_multi_object_tracking\\utils\\utils.py": 1696497789.3935459, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_multi_object_tracking\\utils\\visualization.py": 1696497789.3945448, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_object_segmentation\\aggregate.py": 1696497789.3955438, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_object_segmentation\\cbam.py": 1696497789.396524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_object_segmentation\\eval_network.py": 1696497789.3975399, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_object_segmentation\\inference_core.py": 1696497789.3985462, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_object_segmentation\\inference_memory_bank.py": 1696497789.3985462, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_object_segmentation\\model.py": 1696497789.4005413, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_object_segmentation\\modules.py": 1696497789.4015446, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_object_segmentation\\mod_resnet.py": 1696497789.3995247, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_object_segmentation\\network.py": 1696497789.4015446, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_panoptic_segmentation\\backbone\\swin_checkpoint.py": 1696497789.4055476, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_panoptic_segmentation\\backbone\\swin_transformer.py": 1696497789.4065242, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_panoptic_segmentation\\head\\kernel_head.py": 1696497789.4085429, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_panoptic_segmentation\\head\\kernel_iter_head.py": 1696497789.4095447, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_panoptic_segmentation\\head\\kernel_update_head.py": 1696497789.4095447, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_panoptic_segmentation\\head\\kernel_updator.py": 1696497789.4105444, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_panoptic_segmentation\\head\\mask.py": 1696497789.4115243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_panoptic_segmentation\\head\\semantic_fpn_wrapper.py": 1696497789.4125285, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_panoptic_segmentation\\head\\track_heads.py": 1696497789.4135282, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_panoptic_segmentation\\neck\\fpn.py": 1696497789.414526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_panoptic_segmentation\\track\\quasi_dense_embed_tracker.py": 1696497789.416524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_panoptic_segmentation\\video_k_net.py": 1696497789.4035392, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_panoptic_segmentation\\visualizer.py": 1696497789.4045439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_single_object_tracking\\config\\ostrack.py": 1696497789.4185467, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_single_object_tracking\\models\\layers\\attn.py": 1696497789.420539, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_single_object_tracking\\models\\layers\\attn_blocks.py": 1696497789.4215245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_single_object_tracking\\models\\layers\\head.py": 1696497789.422541, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_single_object_tracking\\models\\layers\\patch_embed.py": 1696497789.422541, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_single_object_tracking\\models\\ostrack\\base_backbone.py": 1696497789.424541, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_single_object_tracking\\models\\ostrack\\ostrack.py": 1696497789.4255292, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_single_object_tracking\\models\\ostrack\\utils.py": 1696497789.4265258, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_single_object_tracking\\models\\ostrack\\vit_ce.py": 1696497789.4265258, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_single_object_tracking\\models\\procontext\\procontext.py": 1696497789.4285526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_single_object_tracking\\models\\procontext\\utils.py": 1696497789.4295409, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_single_object_tracking\\models\\procontext\\vit_ce.py": 1696497789.4305437, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_single_object_tracking\\tracker\\ostrack.py": 1696497789.4315243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_single_object_tracking\\tracker\\procontext.py": 1696497789.4325435, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_single_object_tracking\\utils\\utils.py": 1696497789.4335406, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\DUT\\config.py": 1696497789.4385252, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\DUT\\DUT_raft.py": 1696497789.4365284, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\DUT\\MotionPro.py": 1696497789.4375408, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\DUT\\RAFT\\corr.py": 1696497789.4415414, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\DUT\\RAFT\\extractor.py": 1696497789.442525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\DUT\\RAFT\\raft.py": 1696497789.4435258, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\DUT\\RAFT\\update.py": 1696497789.4445486, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\DUT\\rf_det_module.py": 1696497789.439524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\DUT\\rf_det_so.py": 1696497789.4405239, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\DUT\\Smoother.py": 1696497789.4375408, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\DUTRAFTStabilizer.py": 1696497789.434524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\utils\\image_utils.py": 1696497789.4505265, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\utils\\IterativeSmooth.py": 1696497789.4445486, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\utils\\math_utils.py": 1696497789.4505265, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\utils\\MedianFilter.py": 1696497789.4455254, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\utils\\ProjectionUtils.py": 1696497789.4465241, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\utils\\RAFTUtils.py": 1696497789.4475276, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_stabilization\\utils\\WarpUtils.py": 1696497789.4485261, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_streaming_perception\\longshortnet\\exp\\longshortnet_base.py": 1696497789.4545243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_streaming_perception\\longshortnet\\longshortnet.py": 1696497789.453526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_streaming_perception\\longshortnet\\models\\dfp_pafpn_long.py": 1696497789.4565437, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_streaming_perception\\longshortnet\\models\\dfp_pafpn_short.py": 1696497789.4565437, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_streaming_perception\\longshortnet\\models\\longshort.py": 1696497789.4575245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_streaming_perception\\longshortnet\\models\\longshort_backbone_neck.py": 1696497789.4585278, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_summarization\\base_model.py": 1696497789.4605248, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_summarization\\kts\\cpd_auto.py": 1696497789.463524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_summarization\\kts\\cpd_nonlin.py": 1696497789.464524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_summarization\\pgl_sum.py": 1696497789.4615245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_summarization\\summarizer.py": 1696497789.4615245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_super_resolution\\basicvsr_net.py": 1696497789.4655242, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_super_resolution\\common.py": 1696497789.4665244, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_super_resolution\\msrresnet_lite_model.py": 1696497789.4675245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_super_resolution\\real_basicvsr_for_video_super_resolution.py": 1696497789.4675245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\video_super_resolution\\real_basicvsr_net.py": 1696497789.4685264, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vidt\\backbone.py": 1696497789.4705253, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vidt\\deformable_transformer.py": 1696497789.4715242, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vidt\\fpn_fusion.py": 1696497789.4715242, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vidt\\head.py": 1696497789.472523, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vidt\\model.py": 1696497789.4735243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\virual_tryon\\sdafnet.py": 1696497789.4755256, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vision_efficient_tuning\\backbone.py": 1696497789.4765246, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vision_efficient_tuning\\head.py": 1696497789.477525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vision_efficient_tuning\\model.py": 1696497789.4785247, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vision_efficient_tuning\\petl.py": 1696497789.479524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vision_efficient_tuning\\timm_helpers.py": 1696497789.4805243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vision_efficient_tuning\\timm_vision_transformer.py": 1696497789.4815235, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vision_efficient_tuning\\timm_weight_init.py": 1696497789.4815235, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vision_efficient_tuning\\vision_efficient_tuning.py": 1696497789.4825234, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vision_middleware\\backbone.py": 1696497789.4845235, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vision_middleware\\head.py": 1696497789.4845235, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vision_middleware\\model.py": 1696497789.4855235, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vision_middleware\\vim.py": 1696497789.4865236, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vop_retrieval\\backbone.py": 1696497789.4875238, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vop_retrieval\\basic_utils.py": 1696497789.488524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vop_retrieval\\model.py": 1696497789.4895236, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vop_retrieval\\model_se.py": 1696497789.4905238, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\cv\\vop_retrieval\\tokenization_clip.py": 1696497789.491524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\clip\\bert_tokenizer.py": 1696497789.4965243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\clip\\configuration_bert.py": 1696497789.4975276, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\clip\\model.py": 1696497789.498525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\clip\\modeling_bert.py": 1696497789.4995239, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\clip_interrogator\\model.py": 1696497789.501524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\diffusion\\diffusion.py": 1696497789.5035253, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\diffusion\\model.py": 1696497789.5035253, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\diffusion\\structbert.py": 1696497789.504524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\diffusion\\tokenizer.py": 1696497789.5055263, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\diffusion\\unet_generator.py": 1696497789.5065246, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\diffusion\\unet_upsampler_1024.py": 1696497789.5075247, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\diffusion\\unet_upsampler_256.py": 1696497789.5085247, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\dpm_solver_pytorch.py": 1696497789.4925258, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\efficient_diffusion_tuning\\control_sd_lora.py": 1696497789.5095243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\efficient_diffusion_tuning\\efficient_stable_diffusion.py": 1696497789.510528, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\efficient_diffusion_tuning\\sd_lora.py": 1696497789.5115244, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\gemm\\gemm_base.py": 1696497789.513527, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\gemm\\gemm_model.py": 1696497789.513527, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\gemm\\tokenizer.py": 1696497789.5145254, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\guided_diffusion\\gaussian_diffusion.py": 1696497789.516525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\guided_diffusion\\respace.py": 1696497789.517524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\guided_diffusion\\script.py": 1696497789.517524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\guided_diffusion\\unet.py": 1696497789.5185244, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\image_to_video\\image_to_video_model.py": 1696497789.5205247, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\image_to_video\\modules\\autoencoder.py": 1696497789.5225255, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\image_to_video\\modules\\embedder.py": 1696497789.5225255, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\image_to_video\\modules\\unet_i2v.py": 1696497789.5235238, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\image_to_video\\utils\\config.py": 1696497789.5255249, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\image_to_video\\utils\\diffusion.py": 1696497789.5265257, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\image_to_video\\utils\\seed.py": 1696497789.5275245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\image_to_video\\utils\\shedule.py": 1696497789.5275245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\image_to_video\\utils\\transforms.py": 1696497789.5285237, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mgeo\\backbone.py": 1696497789.5305243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mgeo\\text_classification.py": 1696497789.531525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mgeo\\text_ranking.py": 1696497789.5325434, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mgeo\\token_classification.py": 1696497789.5335438, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mmr\\dataloaders\\rawvideo_util.py": 1696497789.5355425, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mmr\\models\\clip_for_mm_video_embedding.py": 1696497789.5375273, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mmr\\models\\dynamic_inverted_softmax.py": 1696497789.5385277, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mmr\\models\\modeling.py": 1696497789.539526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mmr\\models\\module_clip.py": 1696497789.540545, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mmr\\models\\module_cross.py": 1696497789.540545, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mmr\\models\\tokenization_clip.py": 1696497789.5415246, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mmr\\models\\until_module.py": 1696497789.542525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mplug\\clip\\clip.py": 1696497789.5485241, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mplug\\configuration_mplug.py": 1696497789.5445454, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mplug\\modeling_mplug.py": 1696497789.5455499, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mplug\\mvit.py": 1696497789.5465436, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mplug\\predictor.py": 1696497789.5465436, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mplug_for_all_tasks.py": 1696497789.493525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mplug_owl\\configuration_mplug_owl.py": 1696497789.5505433, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\mplug_owl\\modeling_mplug_owl.py": 1696497789.5515418, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\multi_stage_diffusion\\clip.py": 1696497789.5525527, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\multi_stage_diffusion\\decoder.py": 1696497789.5535243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\multi_stage_diffusion\\gaussian_diffusion.py": 1696497789.5545242, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\multi_stage_diffusion\\model.py": 1696497789.5555258, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\multi_stage_diffusion\\prior.py": 1696497789.5565264, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\multi_stage_diffusion\\tokenizer.py": 1696497789.5575252, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\multi_stage_diffusion\\upsampler.py": 1696497789.5585258, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\multi_stage_diffusion\\xglm.py": 1696497789.5585258, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\configuration_mmspeech.py": 1696497789.5605254, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\configuration_ofa.py": 1696497789.5615263, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\generate\\incremental_decoding_utils.py": 1696497789.568544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\generate\\multihead_attention.py": 1696497789.5695243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\generate\\ngram_repeat_block.py": 1696497789.570539, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\generate\\search.py": 1696497789.5715258, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\generate\\sequence_generator.py": 1696497789.5725276, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\generate\\token_generation_constraints.py": 1696497789.5735545, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\generate\\utils.py": 1696497789.5745416, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\modeling_mmspeech.py": 1696497789.562527, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\modeling_ofa.py": 1696497789.563524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\resnet.py": 1696497789.5645278, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\tokenization_ofa.py": 1696497789.5655274, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\tokenization_ofa_fast.py": 1696497789.5655274, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\utils\\constant.py": 1696497789.5755246, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\utils\\utils.py": 1696497789.5765257, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa\\vit.py": 1696497789.5665274, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa_for_all_tasks.py": 1696497789.494524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\ofa_for_text_to_image_synthesis_model.py": 1696497789.495523, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\prost\\dataloaders\\rawvideo_util.py": 1696497789.5785456, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\prost\\models\\modeling.py": 1696497789.580543, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\prost\\models\\module_clip.py": 1696497789.5815434, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\prost\\models\\module_cross.py": 1696497789.5815434, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\prost\\models\\prost_model.py": 1696497789.582524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\prost\\models\\tokenization_clip.py": 1696497789.5835395, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\prost\\models\\until_config.py": 1696497789.584527, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\prost\\models\\until_module.py": 1696497789.58554, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\rleg\\model.py": 1696497789.5865245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\rleg\\rleg.py": 1696497789.5875401, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\soonet\\blocks.py": 1696497789.5885527, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\soonet\\clip.py": 1696497789.589524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\soonet\\model.py": 1696497789.590539, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\soonet\\swin_transformer.py": 1696497789.590539, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\soonet\\tokenizer.py": 1696497789.5915246, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\soonet\\utils.py": 1696497789.5925252, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\stable_diffusion\\stable_diffusion.py": 1696497789.5945454, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\stable_diffusion\\stable_diffusion_xl.py": 1696497789.595541, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\team\\team_model.py": 1696497789.596524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\team\\utils.py": 1696497789.5975397, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\annotator\\histogram\\palette.py": 1696497789.6235287, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\annotator\\sketch\\pidinet.py": 1696497789.6255252, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\annotator\\sketch\\sketch_simplification.py": 1696497789.626544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\annotator\\util.py": 1696497789.620543, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\autoencoder.py": 1696497789.6135397, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\clip.py": 1696497789.614525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\config.py": 1696497789.614525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\data\\samplers.py": 1696497789.6275249, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\data\\tokenizers.py": 1696497789.628528, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\data\\transforms.py": 1696497789.629528, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\diffusion.py": 1696497789.6155283, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\dpm_solver.py": 1696497789.616524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\mha_flash.py": 1696497789.6175237, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\models\\clip.py": 1696497789.631525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\models\\midas.py": 1696497789.631525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\ops\\degration.py": 1696497789.6335435, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\ops\\distributed.py": 1696497789.6345425, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\ops\\losses.py": 1696497789.6345425, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\ops\\random_mask.py": 1696497789.635525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\ops\\utils.py": 1696497789.6365392, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\unet_sd.py": 1696497789.618527, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\utils\\config.py": 1696497789.6375434, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\utils\\distributed.py": 1696497789.6385274, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\utils\\utils.py": 1696497789.6395268, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\videocomposer\\videocomposer_model.py": 1696497789.6195438, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\video_synthesis\\autoencoder.py": 1696497789.5985458, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\video_synthesis\\diffusion.py": 1696497789.5995247, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\video_synthesis\\text_to_video_synthesis_model.py": 1696497789.600544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\video_synthesis\\unet_sd.py": 1696497789.601526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\video_to_video\\modules\\autoencoder.py": 1696497789.604544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\video_to_video\\modules\\embedder.py": 1696497789.6055248, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\video_to_video\\modules\\unet_v2v.py": 1696497789.6065407, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\video_to_video\\utils\\config.py": 1696497789.6085448, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\video_to_video\\utils\\diffusion_sdedit.py": 1696497789.6085448, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\video_to_video\\utils\\schedules_sdedit.py": 1696497789.6095257, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\video_to_video\\utils\\seed.py": 1696497789.6105242, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\video_to_video\\utils\\solvers_sdedit.py": 1696497789.6115413, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\video_to_video\\utils\\transforms.py": 1696497789.6115413, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\video_to_video\\video_to_video_model.py": 1696497789.6035452, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\vldoc\\convnext.py": 1696497789.6415577, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\vldoc\\conv_fpn_trans.py": 1696497789.6415577, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\vldoc\\model.py": 1696497789.6425252, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\vldoc\\modeling_layout_roberta.py": 1696497789.643541, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\vldoc\\processing.py": 1696497789.644545, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\vldoc\\tokenization.py": 1696497789.6455276, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\multi_modal\\vldoc\\transformer_local.py": 1696497789.6455276, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\bart\\text_error_correction.py": 1696497789.6515245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\bert\\backbone.py": 1696497789.6535425, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\bert\\configuration.py": 1696497789.6535425, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\bert\\document_segmentation.py": 1696497789.654524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\bert\\fill_mask.py": 1696497789.6555414, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\bert\\sentence_embedding.py": 1696497789.6565442, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\bert\\siamese_uie.py": 1696497789.6565442, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\bert\\text_classification.py": 1696497789.6575246, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\bert\\text_ranking.py": 1696497789.6585405, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\bert\\token_classification.py": 1696497789.6585405, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\bert\\word_alignment.py": 1696497789.6595488, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\bloom\\backbone.py": 1696497789.6615474, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\bloom\\text_generation.py": 1696497789.6615474, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\canmt\\canmt_model.py": 1696497789.6635244, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\canmt\\canmt_translation.py": 1696497789.6645267, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\canmt\\sequence_generator.py": 1696497789.6655257, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\chatglm\\configuration.py": 1696497789.6675265, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\chatglm\\quantization.py": 1696497789.6675265, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\chatglm\\text_generation.py": 1696497789.6685264, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\chatglm\\tokenization.py": 1696497789.669526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\chatglm2\\configuration.py": 1696497789.671526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\chatglm2\\quantization.py": 1696497789.6725256, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\chatglm2\\text_generation.py": 1696497789.673525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\chatglm2\\tokenization.py": 1696497789.673525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\codegeex\\codegeex.py": 1696497789.6755242, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\codegeex\\codegeex_for_code_generation.py": 1696497789.676525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\codegeex\\codegeex_for_code_translation.py": 1696497789.6775465, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\codegeex\\inference.py": 1696497789.6785262, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\codegeex\\tokenizer.py": 1696497789.679544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\csanmt\\translation.py": 1696497789.680524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\deberta_v2\\backbone.py": 1696497789.6825235, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\deberta_v2\\configuration.py": 1696497789.6835425, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\deberta_v2\\fill_mask.py": 1696497789.6845257, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\deberta_v2\\tokenization.py": 1696497789.6845257, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\deberta_v2\\tokenization_fast.py": 1696497789.6855235, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\dgds\\backbone.py": 1696497789.6875446, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\dgds\\document_grounded_dialog_generate.py": 1696497789.6875446, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\dgds\\document_grounded_dialog_rerank.py": 1696497789.6885254, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\dgds\\document_grounded_dialog_retrieval.py": 1696497789.689524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\fid_plug\\backbone.py": 1696497789.6925254, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\fid_plug\\configuration.py": 1696497789.6935275, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\fid_plug\\text_generation.py": 1696497789.6945438, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\fid_T5\\text_generation.py": 1696497789.6905243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\glm_130b\\generation\\strategies.py": 1696497789.6985438, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\glm_130b\\initialize.py": 1696497789.6955245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\glm_130b\\quantization\\functional.py": 1696497789.7005396, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\glm_130b\\quantization\\layers.py": 1696497789.7015438, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\glm_130b\\text_generation.py": 1696497789.6965237, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt2\\backbone.py": 1696497789.7025404, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt3\\backbone.py": 1696497789.7045393, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt3\\configuration.py": 1696497789.7045393, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt3\\distributed_gpt3.py": 1696497789.7055497, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt3\\text_generation.py": 1696497789.7065244, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt3\\tokenizer.py": 1696497789.707542, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt_moe\\backbone.py": 1696497789.7095437, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt_moe\\checkpointing.py": 1696497789.7105453, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt_moe\\configuration.py": 1696497789.7105453, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt_moe\\distributed_gpt_moe.py": 1696497789.711524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt_moe\\moe\\experts.py": 1696497789.7145398, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt_moe\\moe\\layer.py": 1696497789.7155244, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt_moe\\moe\\mappings.py": 1696497789.716541, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt_moe\\moe\\sharded_moe.py": 1696497789.716541, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt_moe\\moe\\utils.py": 1696497789.7175243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt_moe\\text_generation.py": 1696497789.7125237, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt_moe\\tokenizer.py": 1696497789.7135246, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\gpt_neo\\backbone.py": 1696497789.7195275, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\heads\\crf_head.py": 1696497789.7205539, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\heads\\fill_mask_head.py": 1696497789.7215304, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\heads\\infromation_extraction_head.py": 1696497789.7225428, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\heads\\text_classification_head.py": 1696497789.723542, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\heads\\text_generation_head.py": 1696497789.723542, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\heads\\text_ranking_head.py": 1696497789.7245283, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\heads\\token_classification_head.py": 1696497789.7255275, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\heads\\torch_pretrain_head.py": 1696497789.7265265, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\hf_transformers\\backbone.py": 1696497789.727527, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\llama\\backbone.py": 1696497789.729524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\llama\\convert_llama_weights_to_hf.py": 1696497789.7305436, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\llama\\text_generation.py": 1696497789.7305436, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\llama2\\backbone.py": 1696497789.7325277, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\llama2\\text_generation.py": 1696497789.7335503, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\lstm\\backbone.py": 1696497789.7345245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\lstm\\token_classification.py": 1696497789.7355394, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\megatron_bert\\backbone.py": 1696497789.7365437, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\megatron_bert\\configuration.py": 1696497789.737524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\megatron_bert\\fill_mask.py": 1696497789.738545, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\arguments.py": 1696497789.7395413, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\blocklm_utils.py": 1696497789.7405248, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\configure_data.py": 1696497789.741524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\data_utils\\corpora.py": 1696497789.7475438, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\data_utils\\datasets.py": 1696497789.7485445, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\data_utils\\extraction.py": 1696497789.7495248, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\data_utils\\file_utils.py": 1696497789.7505424, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\data_utils\\lazy_loader.py": 1696497789.751546, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\data_utils\\samplers.py": 1696497789.7525494, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\data_utils\\sp_tokenizer.py": 1696497789.7525494, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\data_utils\\tokenization.py": 1696497789.7535279, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\data_utils\\tokenization_gpt2.py": 1696497789.7545269, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\data_utils\\wordpiece.py": 1696497789.755526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\generation_utils.py": 1696497789.7425241, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\mglm_for_text_summarization.py": 1696497789.743528, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\model\\distributed.py": 1696497789.7575433, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\model\\downstream.py": 1696497789.7585256, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\model\\modeling_bert.py": 1696497789.7595437, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\model\\modeling_glm.py": 1696497789.7595437, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\model\\prompt.py": 1696497789.760525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\model\\transformer.py": 1696497789.7615237, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\process_grid.py": 1696497789.743528, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\run_test.py": 1696497789.744528, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\test\\test_block.py": 1696497789.7635324, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\test\\test_rel_shift.py": 1696497789.7635324, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\train_utils.py": 1696497789.745527, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\mglm\\utils.py": 1696497789.7465444, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\palm_v2\\configuration.py": 1696497789.7655244, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\palm_v2\\dureader_eval.py": 1696497789.766544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\palm_v2\\text_generation.py": 1696497789.767543, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\peer\\backbone.py": 1696497789.769543, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\peer\\configuration.py": 1696497789.769543, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\peer\\sas_utils.py": 1696497789.7705238, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\peer\\text_classification.py": 1696497789.771544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\plug\\AnnealingLR.py": 1696497789.771544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\plug\\backbone.py": 1696497789.7735255, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\plug\\configuration.py": 1696497789.7745242, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\plug\\distributed_plug.py": 1696497789.7755287, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\plug\\generator.py": 1696497789.776526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\plug_mental\\adv_utils.py": 1696497789.7785263, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\plug_mental\\backbone.py": 1696497789.779526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\plug_mental\\configuration.py": 1696497789.7805266, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\plug_mental\\text_classification.py": 1696497789.7805266, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\polylm\\text_generation.py": 1696497789.7825265, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\ponet\\backbone.py": 1696497789.7845976, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\ponet\\configuration.py": 1696497789.7845976, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\ponet\\document_segmentation.py": 1696497789.7855244, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\ponet\\fill_mask.py": 1696497789.7866042, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\ponet\\tokenization.py": 1696497789.7875278, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\qwen\\backbone.py": 1696497789.7885444, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\qwen\\configuration.py": 1696497789.789542, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\qwen\\qwen_generation_utils.py": 1696497789.7905445, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\qwen\\text_generation.py": 1696497789.791542, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\qwen\\tokenization.py": 1696497789.7925253, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space\\configuration.py": 1696497789.7935247, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space\\dialog_intent_prediction.py": 1696497789.7945254, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space\\dialog_modeling.py": 1696497789.7955465, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space\\dialog_state_tracking.py": 1696497789.7965415, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space\\model\\generator.py": 1696497789.798544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space\\model\\gen_unified_transformer.py": 1696497789.7975237, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space\\model\\intent_unified_transformer.py": 1696497789.799526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space\\model\\model_base.py": 1696497789.799526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space\\model\\tokenization_space.py": 1696497789.800524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space\\model\\unified_transformer.py": 1696497789.8015237, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space\\modules\\embedder.py": 1696497789.802543, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space\\modules\\feedforward.py": 1696497789.8035243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space\\modules\\functions.py": 1696497789.8045413, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space\\modules\\multihead_attention.py": 1696497789.8045413, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space\\modules\\transformer_block.py": 1696497789.805524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space_T_cn\\backbone.py": 1696497789.8075285, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space_T_cn\\configuration.py": 1696497789.8085282, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space_T_cn\\table_question_answering.py": 1696497789.809526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\space_T_en\\text_to_sql.py": 1696497789.8105457, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\structbert\\adv_utils.py": 1696497789.812547, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\structbert\\backbone.py": 1696497789.8135414, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\structbert\\configuration.py": 1696497789.8135414, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\structbert\\faq_question_answering.py": 1696497789.8145435, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\structbert\\fill_mask.py": 1696497789.8155248, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\structbert\\text_classification.py": 1696497789.8165429, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\structbert\\token_classification.py": 1696497789.8175447, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\T5\\backbone.py": 1696497789.6485515, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\T5\\configuration.py": 1696497789.6495254, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\T5\\text2text_generation.py": 1696497789.6495254, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\task_models\\feature_extraction.py": 1696497789.8185318, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\task_models\\fill_mask.py": 1696497789.8195243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\task_models\\information_extraction.py": 1696497789.820544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\task_models\\machine_reading_comprehension.py": 1696497789.8215425, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\task_models\\task_model.py": 1696497789.8215425, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\task_models\\text_classification.py": 1696497789.8225439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\task_models\\text_generation.py": 1696497789.8235242, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\task_models\\text_ranking.py": 1696497789.824524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\task_models\\token_classification.py": 1696497789.8255246, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\unite\\configuration.py": 1696497789.8275447, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\unite\\translation_evaluation.py": 1696497789.828542, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\use\\transformer.py": 1696497789.8295417, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\use\\user_satisfaction_estimation.py": 1696497789.8305254, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\veco\\backbone.py": 1696497789.8325438, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\veco\\configuration.py": 1696497789.833527, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\veco\\fill_mask.py": 1696497789.833527, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\veco\\text_classification.py": 1696497789.8345282, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\veco\\token_classification.py": 1696497789.8355262, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\xlm_roberta\\backbone.py": 1696497789.837525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\nlp\\xlm_roberta\\configuration.py": 1696497789.838545, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\config.py": 1696497789.8405244, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\data\\data_ops.py": 1696497789.843544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\data\\msa_pairing.py": 1696497789.8445463, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\data\\process.py": 1696497789.8455245, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\data\\process_multimer.py": 1696497789.8465412, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\data\\protein.py": 1696497789.8475287, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\data\\residue_constants.py": 1696497789.8475287, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\data\\utils.py": 1696497789.848552, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\dataset.py": 1696497789.8415248, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\model.py": 1696497789.842525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\modules\\alphafold.py": 1696497789.850546, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\modules\\attentions.py": 1696497789.850546, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\modules\\auxillary_heads.py": 1696497789.851524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\modules\\common.py": 1696497789.8525407, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\modules\\confidence.py": 1696497789.8535414, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\modules\\embedders.py": 1696497789.8535414, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\modules\\evoformer.py": 1696497789.8545241, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\modules\\featurization.py": 1696497789.8555262, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\modules\\frame.py": 1696497789.8555262, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\modules\\structure_module.py": 1696497789.8565261, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\modules\\template.py": 1696497789.8575237, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\modules\\triangle_multiplication.py": 1696497789.8585432, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\msa\\mmcif.py": 1696497789.860541, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\msa\\msa_identifiers.py": 1696497789.860541, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\msa\\parsers.py": 1696497789.8615253, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\msa\\pipeline.py": 1696497789.862525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\msa\\templates.py": 1696497789.8635252, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\msa\\tools\\hhblits.py": 1696497789.8655293, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\msa\\tools\\hhsearch.py": 1696497789.8665278, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\msa\\tools\\hmmbuild.py": 1696497789.8675275, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\msa\\tools\\hmmsearch.py": 1696497789.8685272, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\msa\\tools\\jackhmmer.py": 1696497789.8685272, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\msa\\tools\\kalign.py": 1696497789.8695257, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\msa\\tools\\utils.py": 1696497789.87054, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\models\\science\\unifold\\msa\\utils.py": 1696497789.8645253, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\accuracy_metric.py": 1696497788.4495995, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\action_detection_evaluator.py": 1696497788.4495995, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\audio_noise_metric.py": 1696497788.4505987, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\base.py": 1696497788.4515984, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\bleu_metric.py": 1696497788.4525993, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\builder.py": 1696497788.4525993, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\ciderD\\ciderD.py": 1696497788.4791162, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\ciderD\\ciderD_scorer.py": 1696497788.4801154, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\image_colorization_metric.py": 1696497788.4545982, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\image_color_enhance_metric.py": 1696497788.4535985, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\image_denoise_metric.py": 1696497788.4545982, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\image_inpainting_metric.py": 1696497788.455598, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\image_instance_segmentation_metric.py": 1696497788.4566014, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\image_portrait_enhancement_metric.py": 1696497788.4571304, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\image_quality_assessment_degradation_metric.py": 1696497788.4581153, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\image_quality_assessment_mos_metric.py": 1696497788.4601152, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\inbatch_recall_metric.py": 1696497788.4611158, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\loss_metric.py": 1696497788.4631152, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\map_metric.py": 1696497788.464117, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\movie_scene_segmentation_metric.py": 1696497788.4651175, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\ned_metric.py": 1696497788.4661162, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\ocr_recognition_metric.py": 1696497788.4671147, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\ppl_metric.py": 1696497788.4681163, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\prediction_saving_wrapper.py": 1696497788.469117, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\referring_video_object_segmentation_metric.py": 1696497788.470119, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\sequence_classification_metric.py": 1696497788.470119, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\text_generation_metric.py": 1696497788.4711192, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\text_ranking_metric.py": 1696497788.4721155, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\token_classification_metric.py": 1696497788.4731164, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\translation_evaluation_metric.py": 1696497788.4741192, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\video_frame_interpolation_metric.py": 1696497788.4751158, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\video_stabilization_metric.py": 1696497788.4751158, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\video_summarization_metric.py": 1696497788.4771173, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\video_super_resolution_metric\\matlab_functions.py": 1696497788.4811163, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\video_super_resolution_metric\\metric_util.py": 1696497788.482116, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\video_super_resolution_metric\\niqe.py": 1696497788.483119, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\metrics\\video_super_resolution_metric\\video_super_resolution_metric.py": 1696497788.4841154, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\ans_dfsmn_pipeline.py": 1696497789.9988818, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\ans_pipeline.py": 1696497789.999881, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\asr_inference_pipeline.py": 1696502225.2564194, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\asr_wenet_inference_pipeline.py": 1696502227.2985218, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\inverse_text_processing_pipeline.py": 1696497790.0018811, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\kws_farfield_pipeline.py": 1696497790.0028808, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\kws_kwsbp_pipeline.py": 1696497790.0038831, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\language_recognition_eres2net_pipeline.py": 1696497790.004884, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\language_recognition_pipeline.py": 1696497790.004884, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\linear_aec_pipeline.py": 1696497790.005882, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\lm_infer_pipeline.py": 1696502226.5694587, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\punctuation_processing_pipeline.py": 1696502225.6801834, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\segmentation_clustering_pipeline.py": 1696497790.0088823, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\separation_pipeline.py": 1696497790.0099008, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\speaker_change_locating_pipeline.py": 1696497790.0099008, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\speaker_diarization_dialogue_detection_pipeline.py": 1696497790.0108833, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\speaker_diarization_pipeline.py": 1696502225.3770406, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\speaker_diarization_semantic_speaker_turn_detection_pipeline.py": 1696497790.012899, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\speaker_verification_eres2net_pipeline.py": 1696497790.013883, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\speaker_verification_light_pipeline.py": 1696497790.013883, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\speaker_verification_pipeline.py": 1696502225.6901798, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\speaker_verification_rdino_pipeline.py": 1696497790.0158834, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\speech_separation_pipeline.py": 1696502226.585458, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\text_to_speech_pipeline.py": 1696497790.0178819, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\timestamp_pipeline.py": 1696502226.1477816, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\audio\\voice_activity_detection_pipeline.py": 1696502225.948782, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\base.py": 1696497789.9938838, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\builder.py": 1696497789.9948823, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\action_detection_pipeline.py": 1696497790.0208814, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\action_recognition_pipeline.py": 1696497790.0218973, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\animal_recognition_pipeline.py": 1696497790.022885, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\arc_face_recognition_pipeline.py": 1696497790.022885, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\bad_image_detecting_pipeline.py": 1696497790.0238824, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\body_2d_keypoints_pipeline.py": 1696497790.0249035, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\body_3d_keypoints_pipeline.py": 1696497790.0258832, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\card_detection_pipeline.py": 1696497790.0258832, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\cmdssl_video_embedding_pipeline.py": 1696497790.0268984, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\content_check_pipeline.py": 1696497790.027882, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\controllable_image_generation_pipeline.py": 1696497790.0288813, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\crowd_counting_pipeline.py": 1696497790.0298994, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ddcolor_image_colorization_pipeline.py": 1696497790.0298994, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ddpm_semantic_segmentation_pipeline.py": 1696497790.0308816, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\face_attribute_recognition_pipeline.py": 1696497790.0318985, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\face_detection_pipeline.py": 1696497790.0318985, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\face_emotion_pipeline.py": 1696497790.032882, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\face_human_hand_detection_pipeline.py": 1696497790.0338979, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\face_image_generation_pipeline.py": 1696497790.0348856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\face_liveness_ir_pipeline.py": 1696497790.0348856, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\face_liveness_xc_pipeline.py": 1696497790.0358834, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\face_processing_base_pipeline.py": 1696497790.0368853, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\face_quality_assessment_pipeline.py": 1696497790.0379057, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\face_recognition_onnx_fm_pipeline.py": 1696497790.0389056, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\face_recognition_onnx_ir_pipeline.py": 1696497790.0389056, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\face_recognition_ood_pipeline.py": 1696497790.0398817, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\face_recognition_pipeline.py": 1696497790.040898, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\face_reconstruction_pipeline.py": 1696497790.041902, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\facial_expression_recognition_pipeline.py": 1696497790.041902, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\facial_landmark_confidence_pipeline.py": 1696497790.0428824, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\fast_instance_segmentation_pipeline.py": 1696497790.043902, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\general_recognition_pipeline.py": 1696497790.0449016, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\hand_static_pipeline.py": 1696497790.0449016, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\hicossl_video_embedding_pipeline.py": 1696497790.0458813, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\human_image_generation_pipeline.py": 1696497790.0468984, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\human_reconstruction_pipeline.py": 1696497790.0468984, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_body_reshaping_pipeline.py": 1696497790.0478866, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_bts_depth_estimation_pipeline.py": 1696497790.0488842, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_cartoon_pipeline.py": 1696497790.0499034, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_classification_pipeline.py": 1696497790.0509014, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_colorization_pipeline.py": 1696497790.0518813, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_color_enhance_pipeline.py": 1696497790.0509014, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_debanding_pipeline.py": 1696497790.0528977, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_deblur_pipeline.py": 1696497790.0528977, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_defrcn_fewshot_pipeline.py": 1696497790.053902, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_denoise_pipeline.py": 1696497790.0548818, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_depth_estimation_pipeline.py": 1696497790.0559032, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_detection_pipeline.py": 1696497790.0559032, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_driving_perception_pipeline.py": 1696497790.0568817, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_editing_pipeline.py": 1696497790.057901, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_face_fusion_pipeline.py": 1696497790.058887, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_human_parsing_pipeline.py": 1696497790.059906, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_inpainting_pipeline.py": 1696497790.059906, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_inpainting_sdv2_pipeline.py": 1696497790.0608819, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_instance_segmentation_pipeline.py": 1696497790.0618992, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_matching_pipeline.py": 1696497790.0629015, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_matting_pipeline.py": 1696497790.0629015, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_mvs_depth_estimation_pipeline.py": 1696497790.0638816, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_open_vocabulary_detection_pipeline.py": 1696497790.0648985, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_paintbyexample_pipeline.py": 1696497790.0648985, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_panoptic_segmentation_pipeline.py": 1696497790.0659075, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_portrait_enhancement_pipeline.py": 1696497790.0669923, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_quality_assessment_degradation_pipeline.py": 1696497790.0669923, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_quality_assessment_man_pipeline.py": 1696497790.0679862, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_quality_assessment_mos_pipeline.py": 1696497790.0689769, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_reid_person_pipeline.py": 1696497790.0699944, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_restoration_pipeline.py": 1696497790.0709755, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_salient_detection_pipeline.py": 1696497790.0709755, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_semantic_segmentation_pipeline.py": 1696497790.0719774, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_skychange_pipeline.py": 1696497790.0729754, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_structured_model_probing_pipeline.py": 1696497790.0739946, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_style_transfer_pipeline.py": 1696497790.074993, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_super_resolution_pasd_pipeline.py": 1696497790.074993, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_super_resolution_pipeline.py": 1696497790.075976, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_to_image_generate_pipeline.py": 1696497790.0769758, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_to_image_translation_pipeline.py": 1696497790.0779963, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\image_try_on_pipeline.py": 1696497790.0779963, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\indoor_layout_estimation_pipeline.py": 1696497790.0789952, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\language_guided_video_summarization_pipeline.py": 1696497790.0799792, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\license_plate_detection_pipeline.py": 1696497790.0809765, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\lineless_table_recognition_pipeline.py": 1696497790.0819914, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\live_category_pipeline.py": 1696497790.0819914, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\maskdino_instance_segmentation_pipeline.py": 1696497790.0839753, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\mask_face_recognition_pipeline.py": 1696497790.0829928, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\mobile_image_super_resolution_pipeline.py": 1696497790.084994, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\mog_face_detection_pipeline.py": 1696497790.084994, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\motion_generation_pipeline.py": 1696497790.0859754, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\movie_scene_segmentation_pipeline.py": 1696497790.086994, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\mtcnn_face_detection_pipeline.py": 1696497790.0879793, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\nerf_recon_4k_pipeline.py": 1696497790.088978, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\nerf_recon_acc_pipeline.py": 1696497790.0904624, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\nerf_recon_vq_compression_pipeline.py": 1696497790.0918722, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\object_detection_3d_pipeline.py": 1696497790.0930114, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ocr_detection_pipeline.py": 1696497790.0941153, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ocr_recognition_pipeline.py": 1696497790.0957036, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ocr_utils\\model_convnext_transformer.py": 1696497790.1330523, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ocr_utils\\model_dla34.py": 1696497790.1340504, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ocr_utils\\model_resnet18_half.py": 1696497790.1340504, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ocr_utils\\model_resnet_mutex_v4_linewithchar.py": 1696497790.1350338, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ocr_utils\\model_vlpt.py": 1696497790.1360488, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ocr_utils\\ocr_modules\\convnext.py": 1696497790.1410496, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ocr_utils\\ocr_modules\\timm_tinyc.py": 1696497790.1420338, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ocr_utils\\ocr_modules\\vitstr.py": 1696497790.1430523, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ocr_utils\\ops.py": 1696497790.137051, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ocr_utils\\resnet18_v1.py": 1696497790.137051, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ocr_utils\\resnet_utils.py": 1696497790.1380339, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ocr_utils\\table_process.py": 1696497790.1390345, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ocr_utils\\utils.py": 1696497790.1400511, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\panorama_depth_estimation_pipeline.py": 1696497790.096229, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\panorama_depth_estimation_s2net_pipeline.py": 1696497790.0972748, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\pedestrian_attribute_recognition_pipeline.py": 1696497790.0977974, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\pointcloud_sceneflow_estimation_pipeline.py": 1696497790.0988572, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\product_retrieval_embedding_pipeline.py": 1696497790.099388, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\product_segmentation_pipeline.py": 1696497790.1004882, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\realtime_video_object_detection_pipeline.py": 1696497790.10104, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\referring_video_object_segmentation_pipeline.py": 1696497790.1020367, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\retina_face_detection_pipeline.py": 1696497790.1030364, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\shop_segmentation_pipleline.py": 1696497790.1040378, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\skin_retouching_pipeline.py": 1696497790.1050375, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\surface_recon_common_pipeline.py": 1696497790.1060348, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\table_recognition_pipeline.py": 1696497790.107034, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\tbs_detection_pipeline.py": 1696497790.108033, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\tbs_detection_utils\\utils.py": 1696497790.1450543, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\text_driven_segmentation_pipleline.py": 1696497790.1090338, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\text_to_360panorama_image_pipeline.py": 1696497790.1100347, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\tinynas_classification_pipeline.py": 1696497790.1100347, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\tinynas_detection_pipeline.py": 1696497790.1120348, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\ulfd_face_detection_pipeline.py": 1696497790.1130345, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\video_category_pipeline.py": 1696497790.1140354, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\video_colorization_pipeline.py": 1696497790.115034, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\video_deinterlace_pipeline.py": 1696497790.1160357, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\video_depth_estimation_pipeline.py": 1696497790.1170352, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\video_frame_interpolation_pipeline.py": 1696497790.1180336, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\video_human_matting_pipeline.py": 1696497790.1190329, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\video_inpainting_pipeline.py": 1696497790.1200328, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\video_instance_segmentation_pipeline.py": 1696497790.121036, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\video_multi_object_tracking_pipeline.py": 1696497790.122034, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\video_object_segmentation_pipeline.py": 1696497790.123035, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\video_panoptic_segmentation_pipeline.py": 1696497790.123035, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\video_single_object_tracking_pipeline.py": 1696497790.124034, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\video_stabilization_pipeline.py": 1696497790.1250336, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\video_summarization_pipeline.py": 1696497790.1250336, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\video_super_resolution_pipeline.py": 1696497790.1260338, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\vidt_pipeline.py": 1696497790.1270354, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\virtual_try_on_pipeline.py": 1696497790.1280365, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\vision_efficient_tuning_pipeline.py": 1696497790.1290545, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\vision_middleware_pipeline.py": 1696497790.1300378, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\vop_retrieval_pipeline.py": 1696497790.1300378, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\cv\\vop_retrieval_se_pipeline.py": 1696497790.1310377, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\asr_pipeline.py": 1696497790.1460335, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\cone2_pipeline\\cones2_inference_pipeline.py": 1696497790.1670525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\diffusers_wrapped\\devices.py": 1696497790.1684558, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\diffusers_wrapped\\diffusers_pipeline.py": 1696497790.169466, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\diffusers_wrapped\\pasd_pipeline.py": 1696497790.1704395, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\diffusers_wrapped\\stable_diffusion\\chinese_stable_diffusion_pipeline.py": 1696497790.173456, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\diffusers_wrapped\\stable_diffusion\\stable_diffusion_pipeline.py": 1696497790.1744578, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\diffusers_wrapped\\vaehook.py": 1696497790.1714394, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\disco_guided_diffusion_pipeline\\disco_guided_diffusion.py": 1696497790.1764402, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\disco_guided_diffusion_pipeline\\utils.py": 1696497790.1774428, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\document_vl_embedding_pipeline.py": 1696497790.1470494, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\efficient_diffusion_tuning_pipeline.py": 1696497790.1480384, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\generative_multi_modal_embedding_pipeline.py": 1696497790.1480384, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\gridvlp_pipeline.py": 1696497790.1490376, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\image_captioning_pipeline.py": 1696497790.1500363, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\image_text_retrieval_pipeline.py": 1696497790.1510527, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\image_to_video_pipeline.py": 1696497790.1510527, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\mgeo_ranking_pipeline.py": 1696497790.1520333, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\multimodal_dialogue_pipeline.py": 1696497790.1530535, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\multi_modal_embedding_pipeline.py": 1696497790.1530535, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\ocr_recognition_pipeline.py": 1696497790.1540344, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\prost_text_video_retrieval_pipeline.py": 1696497790.155033, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\soonet_video_temporal_grounding_pipeline.py": 1696497790.1560533, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\sudoku_pipeline.py": 1696497790.1560533, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\team_multi_modal_similarity_pipeline.py": 1696497790.1570342, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\text2sql_pipeline.py": 1696497790.1580346, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\text_to_image_synthesis_pipeline.py": 1696497790.1590345, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\text_to_video_synthesis_pipeline.py": 1696497790.1600525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\videocomposer_pipeline.py": 1696497790.1630514, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\video_captioning_pipeline.py": 1696497790.1600525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\video_multi_modal_embedding_pipeline.py": 1696497790.1610348, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\video_question_answering_pipeline.py": 1696497790.1620524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\video_to_video_pipeline.py": 1696497790.1630514, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\visual_entailment_pipeline.py": 1696497790.1640341, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\visual_grounding_pipeline.py": 1696497790.1650333, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\multi_modal\\visual_question_answering_pipeline.py": 1696497790.1660483, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\automatic_post_editing_pipeline.py": 1696497790.1794567, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\canmt_translation_pipeline.py": 1696497790.1794567, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\codegeex_code_generation_pipeline.py": 1696497790.1804583, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\codegeex_code_translation_pipeline.py": 1696497790.1814392, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\conversational_text_to_sql_pipeline.py": 1696497790.1824594, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\dialog_intent_prediction_pipeline.py": 1696497790.1824594, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\dialog_modeling_pipeline.py": 1696497790.1834397, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\dialog_state_tracking_pipeline.py": 1696497790.184457, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\distributed_gpt3_pipeline.py": 1696497790.184457, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\distributed_gpt_moe_pipeline.py": 1696497790.1854587, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\distributed_plug_pipeline.py": 1696497790.1864398, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\document_grounded_dialog_generate_pipeline.py": 1696497790.1864398, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\document_grounded_dialog_rerank_pipeline.py": 1696497790.187455, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\document_grounded_dialog_retrieval_pipeline.py": 1696497790.1884432, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\document_segmentation_pipeline.py": 1696497790.1894407, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\extractive_summarization_pipeline.py": 1696497790.1904557, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\faq_question_answering_pipeline.py": 1696497790.1914592, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\fasttext_text_classification_pipeline.py": 1696497790.1924407, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\feature_extraction_pipeline.py": 1696497790.1924407, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\fid_dialogue_pipeline.py": 1696497790.1934407, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\fill_mask_pipeline.py": 1696497790.1944413, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\glm130b_text_generation_pipeline.py": 1696497790.1954546, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\information_extraction_pipeline.py": 1696497790.1954546, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\interactive_translation_pipeline.py": 1696497790.1964395, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\language_identification_pipline.py": 1696497790.1974382, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\machine_reading_comprehension_pipeline.py": 1696497790.1984394, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\mglm_text_summarization_pipeline.py": 1696497790.1994562, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\named_entity_recognition_pipeline.py": 1696497790.1994562, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\polylm_text_generation_pipeline.py": 1696497790.2004397, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\sentence_embedding_pipeline.py": 1696497790.2014556, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\siamese_uie_pipeline.py": 1696497790.2014556, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\summarization_pipeline.py": 1696497790.20244, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\table_question_answering_pipeline.py": 1696497790.203439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\text_classification_pipeline.py": 1696497790.2044542, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\text_error_correction_pipeline.py": 1696497790.2044542, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\text_generation_pipeline.py": 1696497790.205439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\text_ranking_pipeline.py": 1696497790.2064548, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\token_classification_pipeline.py": 1696497790.2064548, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\translation_evaluation_pipeline.py": 1696497790.2074575, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\translation_pipeline.py": 1696497790.2084403, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\translation_quality_estimation_pipeline.py": 1696497790.2094405, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\user_satisfaction_estimation_pipeline.py": 1696497790.2104387, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\word_alignment_pipeline.py": 1696497790.2114568, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\word_segmentation_pipeline.py": 1696497790.2114568, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\nlp\\zero_shot_classification_pipeline.py": 1696497790.212439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\pipeline_template.py": 1696497789.995881, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\science\\protein_structure_pipeline.py": 1696497790.2144585, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\pipelines\\util.py": 1696497789.996881, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\asr.py": 1696497790.2154424, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\audio.py": 1696497790.2164404, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\base.py": 1696497790.2174413, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\builder.py": 1696497790.21844, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\common.py": 1696497790.21844, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\cv\\action_detection_mapper.py": 1696497790.2254405, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\cv\\bad_image_detecting_preprocessor.py": 1696497790.226441, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\cv\\controllable_image_generation.py": 1696497790.227456, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\cv\\cv2_transforms.py": 1696497790.227456, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\cv\\image_classification_preprocessor.py": 1696497790.2284396, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\cv\\image_quality_assessment_man.py": 1696497790.2294388, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\cv\\image_quality_assessment_mos.py": 1696497790.2304592, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\cv\\image_restoration_preprocessor.py": 1696497790.2314427, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\cv\\mmcls_preprocessor.py": 1696497790.2314427, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\cv\\timer.py": 1696497790.2324393, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\cv\\util.py": 1696497790.233439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\cv\\video_stabilization.py": 1696497790.2344553, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\cv\\video_super_resolution.py": 1696497790.2344553, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\image.py": 1696497790.2194414, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\kws.py": 1696497790.220441, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\movie_scene_segmentation\\transforms.py": 1696497790.2364385, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\multi_modal.py": 1696497790.2214406, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\bert_seq_cls_tokenizer.py": 1696497790.2374554, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\canmt_translation.py": 1696497790.2384567, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\dialog_classification_use_preprocessor.py": 1696497790.2394395, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\document_grounded_dialog_generate_preprocessor.py": 1696497790.2404587, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\document_grounded_dialog_rerank_preprocessor.py": 1696497790.2404587, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\document_grounded_dialog_retrieval_preprocessor.py": 1696497790.2414443, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\document_segmentation_preprocessor.py": 1696497790.2424407, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\faq_question_answering_preprocessor.py": 1696497790.2434592, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\feature_extraction_preprocessor.py": 1696497790.2444398, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\fill_mask_preprocessor.py": 1696497790.2444398, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\machine_reading_comprehension_preprocessor.py": 1696497790.2454433, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\mgeo_ranking_preprocessor.py": 1696497790.2464392, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\mglm_summarization_preprocessor.py": 1696497790.2474568, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\relation_extraction_preprocessor.py": 1696497790.2474568, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\sentence_embedding_preprocessor.py": 1696497790.2484393, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\siamese_uie_preprocessor.py": 1696497790.2494586, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space\\args.py": 1696497790.2614594, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space\\batch.py": 1696497790.2614594, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space\\data_loader.py": 1696497790.262439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space\\dialog_intent_prediction_preprocessor.py": 1696497790.2634387, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space\\dialog_modeling_preprocessor.py": 1696497790.2644563, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space\\dialog_state_tracking_preprocessor.py": 1696497790.2644563, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space\\dst_processors.py": 1696497790.2654393, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space\\fields\\gen_field.py": 1696497790.2704382, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space\\fields\\intent_field.py": 1696497790.2724423, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space\\lazy_dataset.py": 1696497790.266439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space\\preprocess.py": 1696497790.267458, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space\\sampler.py": 1696497790.267458, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space\\tensorlistdataset.py": 1696497790.2684402, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space\\tokenizer.py": 1696497790.2694397, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space_T_cn\\fields\\database.py": 1696497790.2754407, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space_T_cn\\fields\\schema_link.py": 1696497790.27644, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space_T_cn\\fields\\struct.py": 1696497790.2774415, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space_T_cn\\table_question_answering_preprocessor.py": 1696497790.2744424, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space_T_en\\conversational_text_to_sql_preprocessor.py": 1696497790.2794416, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space_T_en\\fields\\common_utils.py": 1696497790.2804668, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space_T_en\\fields\\parse.py": 1696497790.2814395, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space_T_en\\fields\\preprocess_dataset.py": 1696497790.2824552, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\space_T_en\\fields\\process_dataset.py": 1696497790.28346, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\text_classification_preprocessor.py": 1696497790.2504683, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\text_clean.py": 1696497790.2504683, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\text_error_correction.py": 1696497790.2514398, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\text_generation_preprocessor.py": 1696497790.2524388, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\text_ranking_preprocessor.py": 1696497790.2534387, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\token_classification_preprocessor.py": 1696497790.2544582, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\token_classification_thai_preprocessor.py": 1696497790.2544582, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\token_classification_viet_preprocessor.py": 1696497790.2554412, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\transformers_tokenizer.py": 1696497790.2564394, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\translation_evaluation_preprocessor.py": 1696497790.2564394, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\utils.py": 1696497790.257459, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\word_alignment_preprocessor.py": 1696497790.2584393, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\nlp\\zero_shot_classification_preprocessor.py": 1696497790.2594419, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\asr.py": 1696497790.2844608, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\base.py": 1696497790.285439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\image_captioning.py": 1696497790.2864594, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\image_classification.py": 1696497790.287442, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\ocr_recognition.py": 1696497790.287442, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\sudoku.py": 1696497790.2884395, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\summarization.py": 1696497790.2894583, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\text2sql.py": 1696497790.2904558, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\text_classification.py": 1696497790.2904558, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\text_to_image_synthesis.py": 1696497790.291439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\utils\\audio_helper.py": 1696497790.295454, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\utils\\bridge_content_encoder.py": 1696497790.295454, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\utils\\collate.py": 1696497790.2964394, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\utils\\constant.py": 1696497790.2974577, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\utils\\get_tables.py": 1696497790.2984433, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\utils\\random_help.py": 1696497790.2984433, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\utils\\text2phone.py": 1696497790.2994566, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\utils\\transforms.py": 1696497790.3004398, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\utils\\vision_helper.py": 1696497790.301456, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\visual_entailment.py": 1696497790.292439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\visual_grounding.py": 1696497790.292439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\ofa\\visual_question_answering.py": 1696497790.293439, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\science\\uni_fold.py": 1696497790.3024578, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\speaker.py": 1696497790.2214406, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\tts.py": 1696497790.2224386, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\preprocessors\\video.py": 1696497790.2234411, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\audio\\ans_trainer.py": 1696497790.314082, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\audio\\asr_trainer.py": 1696502225.3460402, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\audio\\kws_farfield_trainer.py": 1696497790.3150704, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\audio\\kws_nearfield_trainer.py": 1696497790.3160667, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\audio\\kws_utils\\batch_utils.py": 1696497790.320088, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\audio\\kws_utils\\det_utils.py": 1696497790.320088, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\audio\\kws_utils\\file_utils.py": 1696497790.321067, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\audio\\kws_utils\\model_utils.py": 1696497790.3220663, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\audio\\kws_utils\\runtime_utils.py": 1696497790.3220663, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\audio\\separation_trainer.py": 1696497790.317069, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\audio\\tts_trainer.py": 1696497790.3180687, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\base.py": 1696497790.3074389, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\builder.py": 1696497790.3074389, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cli_argument_parser.py": 1696497790.3084586, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\action_detection_trainer.py": 1696497790.3240664, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\card_detection_scrfd_trainer.py": 1696497790.3250823, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\cartoon_translation_trainer.py": 1696497790.3260877, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\face_detection_scrfd_trainer.py": 1696497790.3270676, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\image_classifition_trainer.py": 1696497790.3270676, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\image_defrcn_fewshot_detection_trainer.py": 1696497790.328067, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\image_detection_damoyolo_trainer.py": 1696497790.3290663, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\image_inpainting_trainer.py": 1696497790.3300664, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\image_instance_segmentation_trainer.py": 1696497790.3300664, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\image_portrait_enhancement_trainer.py": 1696497790.3310688, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\movie_scene_segmentation_trainer.py": 1696497790.3320684, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\nerf_recon_acc_trainer.py": 1696497790.3330672, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\ocr_detection_db_trainer.py": 1696497790.3340676, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\ocr_recognition_trainer.py": 1696497790.3340676, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\referring_video_object_segmentation_trainer.py": 1696497790.3350668, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\cv\\vision_efficient_tuning_trainer.py": 1696497790.336067, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\default_config.py": 1696497790.3094437, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\builder.py": 1696497790.3380673, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\checkpoint\\checkpoint_hook.py": 1696497790.3440847, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\checkpoint\\checkpoint_processor.py": 1696497790.345067, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\checkpoint\\load_checkpoint_hook.py": 1696497790.3460834, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\clip_clamp_logit_scale_hook.py": 1696497790.3380673, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\compression\\sparsity_hook.py": 1696497790.347086, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\compression\\utils.py": 1696497790.3480673, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\distributed\\ddp_hook.py": 1696497790.350082, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\distributed\\deepspeed_hook.py": 1696497790.350082, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\distributed\\megatron_hook.py": 1696497790.351067, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\early_stop_hook.py": 1696497790.339069, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\evaluation_hook.py": 1696497790.3400865, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\hook.py": 1696497790.3410673, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\iter_timer_hook.py": 1696497790.3410673, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\logger\\base.py": 1696497790.3530686, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\logger\\tensorboard_hook.py": 1696497790.3540742, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\logger\\text_logger_hook.py": 1696497790.3540742, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\lr_scheduler_hook.py": 1696497790.34207, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\optimizer\\apex_optimizer_hook.py": 1696497790.3560686, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\optimizer\\base.py": 1696497790.3570683, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\optimizer\\torch_optimizer_hook.py": 1696497790.3590684, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\hooks\\priority.py": 1696497790.3430872, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\lrscheduler\\builder.py": 1696497790.3610718, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\lrscheduler\\warmup\\base.py": 1696497790.36207, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\lrscheduler\\warmup\\warmup.py": 1696497790.3630707, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\multi_modal\\clip\\clip_trainer.py": 1696497790.3670697, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\multi_modal\\clip\\clip_trainer_utils.py": 1696497790.368068, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\multi_modal\\cones2\\cones_trainer.py": 1696497790.3700695, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\multi_modal\\custom_diffusion\\custom_diffusion_trainer.py": 1696497790.3720665, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\multi_modal\\dreambooth_diffusion\\dreambooth_diffusion_trainer.py": 1696497790.3730671, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\multi_modal\\efficient_diffusion_tuning\\efficient_diffusion_tuning_trainer.py": 1696497790.3750837, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\multi_modal\\lora_diffusion\\lora_diffusion_trainer.py": 1696497790.376067, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\multi_modal\\lora_diffusion_xl\\lora_diffusion_xl_trainer.py": 1696497790.3780675, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\multi_modal\\mgeo_ranking_trainer.py": 1696497790.3650696, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\multi_modal\\mplug\\mplug_trainer.py": 1696497790.3800676, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\multi_modal\\ofa\\ofa_trainer.py": 1696497790.3820906, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\multi_modal\\ofa\\ofa_trainer_utils.py": 1696497790.3820906, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\multi_modal\\stable_diffusion\\stable_diffusion_trainer.py": 1696497790.3840666, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\multi_modal\\team\\team_trainer.py": 1696497790.3850858, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\multi_modal\\team\\team_trainer_utils.py": 1696497790.386067, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\csanmt_translation_trainer.py": 1696497790.3880672, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\document_grounded_dialog_generate_trainer.py": 1696497790.3880672, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\document_grounded_dialog_rerank_trainer.py": 1696497790.3890865, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\document_grounded_dialog_retrieval_trainer.py": 1696497790.3900673, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\faq_question_answering_trainer.py": 1696497790.3910835, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\gpt3_trainer.py": 1696497790.3910835, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\gpt_moe_trainer.py": 1696497790.3920825, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\plug_trainer.py": 1696497790.3930707, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\sentence_embedding_trainer.py": 1696497790.3940701, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\sequence_classification_trainer.py": 1696497790.3950703, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\siamese_uie_trainer.py": 1696497790.3950703, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\space\\dialog_intent_trainer.py": 1696497790.400071, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\space\\dialog_modeling_trainer.py": 1696497790.4010694, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\space\\eval.py": 1696497790.4020855, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\space\\metrics\\metrics_tracker.py": 1696497790.4030867, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\space\\trainer\\gen_trainer.py": 1696497790.404086, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\space\\trainer\\intent_trainer.py": 1696497790.4050932, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\table_question_answering_trainer.py": 1696497790.3960676, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\text_generation_trainer.py": 1696497790.3970666, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\text_ranking_trainer.py": 1696497790.398067, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp\\translation_evaluation_trainer.py": 1696497790.398067, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\nlp_trainer.py": 1696497790.3105536, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\optimizer\\builder.py": 1696497790.4070835, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\optimizer\\child_tuning_adamw_optimizer.py": 1696497790.4080677, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\parallel\\builder.py": 1696497790.4090672, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\parallel\\utils.py": 1696497790.410067, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\trainer.py": 1696497790.311074, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\training_args.py": 1696497790.312067, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\utils\\inference.py": 1696497790.412069, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\trainers\\utils\\log_buffer.py": 1696497790.412069, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\audio\\asr_dataset.py": 1696497789.8735435, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\auth\\auth_config.py": 1696497789.875524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\context\\dataset_context_config.py": 1696497789.8765464, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\audio\\asr_dataset.py": 1696497789.8945236, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\audio\\kws_farfield_dataset.py": 1696497789.895541, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\audio\\kws_nearfield_dataset.py": 1696497789.895541, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\audio\\kws_nearfield_processor.py": 1696497789.896524, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\bad_image_detecting\\bad_image_detecting_dataset.py": 1696497789.8985438, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\builder.py": 1696497789.883526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\damoyolo\\build.py": 1696497789.900544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\damoyolo\\collate_batch.py": 1696497789.900544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\damoyolo\\datasets\\coco.py": 1696497789.9025538, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\damoyolo\\datasets\\mosaic_wrapper.py": 1696497789.903069, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\damoyolo\\evaluation\\coco\\coco_eval.py": 1696497789.9060805, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\damoyolo\\samplers\\distributed.py": 1696497789.9070675, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\damoyolo\\samplers\\grouped_batch_sampler.py": 1696497789.9080634, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\damoyolo\\samplers\\iteration_based_batch_sampler.py": 1696497789.9090638, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\damoyolo\\transforms\\build.py": 1696497789.911066, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\damoyolo\\transforms\\transforms.py": 1696497789.911066, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\easycv_base.py": 1696497789.8845236, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\gopro_image_deblurring_dataset.py": 1696497789.8855243, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\image_colorization\\image_colorization_dataset.py": 1696497789.9130628, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\image_inpainting\\aug.py": 1696497789.9140804, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\image_inpainting\\image_inpainting_dataset.py": 1696497789.9150803, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\image_instance_segmentation_coco_dataset.py": 1696497789.8865242, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\image_portrait_enhancement\\data_utils.py": 1696497789.9170816, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\image_portrait_enhancement\\image_portrait_enhancement_dataset.py": 1696497789.9180624, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\image_quality_assessment_degradation\\image_quality_assessment_degradation_dataset.py": 1696497789.919083, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\image_quality_assmessment_mos\\image_quality_assessment_mos_dataset.py": 1696497789.9210627, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\language_guided_video_summarization_dataset.py": 1696497789.8865242, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\mgeo_ranking_dataset.py": 1696497789.8875246, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\movie_scene_segmentation\\movie_scene_segmentation_dataset.py": 1696497789.9230807, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\movie_scene_segmentation\\sampler.py": 1696497789.9230807, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\ocr_detection\\augmenter.py": 1696497789.9250796, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\ocr_detection\\data_loader.py": 1696497789.9250796, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\ocr_detection\\image_dataset.py": 1696497789.9260676, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\ocr_detection\\measures\\iou_evaluator.py": 1696497789.9280622, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\ocr_detection\\measures\\quad_measurer.py": 1696497789.9290822, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\ocr_detection\\processes\\augment_data.py": 1696497789.9300847, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\ocr_detection\\processes\\data_process.py": 1696497789.931063, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\ocr_detection\\processes\\make_border_map.py": 1696497789.93208, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\ocr_detection\\processes\\make_icdar_data.py": 1696497789.9330828, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\ocr_detection\\processes\\make_seg_detection_data.py": 1696497789.9330828, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\ocr_detection\\processes\\normalize_image.py": 1696497789.934063, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\ocr_detection\\processes\\random_crop_data.py": 1696497789.935081, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\ocr_recognition_dataset.py": 1696497789.888528, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\reds_image_deblurring_dataset.py": 1696497789.889527, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\referring_video_object_segmentation\\referring_video_object_segmentation_dataset.py": 1696497789.9360802, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\referring_video_object_segmentation\\transformers.py": 1696497789.9370623, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\sidd_image_denoising\\data_utils.py": 1696497789.9390795, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\sidd_image_denoising\\sidd_image_denoising_dataset.py": 1696497789.9400842, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\sidd_image_denoising\\transforms.py": 1696497789.9400842, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\text_ranking_dataset.py": 1696497789.890544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\torch_custom_dataset.py": 1696497789.8915238, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\veco_dataset.py": 1696497789.892525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\video_frame_interpolation\\data_utils.py": 1696497789.9420626, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\video_frame_interpolation\\video_frame_interpolation_dataset.py": 1696497789.943064, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\video_stabilization\\video_stabilization_dataset.py": 1696497789.945084, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\video_summarization_dataset.py": 1696497789.892525, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\custom_datasets\\video_super_resolution\\video_super_resolution_dataset.py": 1696497789.9460628, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\dataset_cls\\dataset.py": 1696497789.8825428, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\data_files\\data_files_manager.py": 1696497789.878544, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\data_loader\\data_loader.py": 1696497789.880526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\data_loader\\data_loader_manager.py": 1696497789.880526, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\download\\dataset_builder.py": 1696497789.9480646, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\download\\download_config.py": 1696497789.9480646, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\download\\download_manager.py": 1696497789.9490626, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\meta\\data_meta_config.py": 1696497789.9510813, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\meta\\data_meta_manager.py": 1696497789.952084, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\ms_dataset.py": 1696497789.8725457, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\task_datasets\\gopro_image_deblurring_dataset.py": 1696497789.9530628, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\task_datasets\\reds_image_deblurring_dataset.py": 1696497789.954083, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\task_datasets\\sidd_image_denoising.py": 1696497789.954083, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\task_datasets\\torch_base_dataset.py": 1696497789.9550626, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\task_datasets\\video_summarization_dataset.py": 1696497789.9560826, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\utils\\dataset_utils.py": 1696497789.9570813, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\utils\\delete_utils.py": 1696497789.9580631, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\utils\\maxcompute_utils.py": 1696497789.959065, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\utils\\oss_utils.py": 1696497789.960081, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\msdatasets\\utils\\upload_utils.py": 1696497789.960081, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\exporters\\audio\\ans_dfsmn_exporter.py": 1696497788.421503, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\exporters\\base.py": 1696497788.417513, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\exporters\\builder.py": 1696497788.4185057, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\exporters\\cv\\cartoon_translation_exporter.py": 1696497788.4235177, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\exporters\\cv\\face_detection_scrfd_exporter.py": 1696497788.4245052, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\exporters\\cv\\object_detection_damoyolo_exporter.py": 1696497788.4245052, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\exporters\\multi_modal\\stable_diffusion_exporter.py": 1696497788.4275067, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\exporters\\nlp\\csanmt_for_translation_exporter.py": 1696497788.428513, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\exporters\\nlp\\model_for_token_classification_exporter.py": 1696497788.4295032, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\exporters\\nlp\\sbert_for_sequence_classification_exporter.py": 1696497788.430504, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\exporters\\nlp\\sbert_for_zero_shot_classification_exporter.py": 1696497788.4315038, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\exporters\\tf_model_exporter.py": 1696497788.4195058, "D:\\PythonProject\\bert-vits2-all\\Bert-VITS2\\env\\Lib\\site-packages\\modelscope\\exporters\\torch_model_exporter.py": 1696497788.4205043}, "modelscope_path": "D:/PythonProject/bert-vits2-all/Bert-VITS2/env/Lib/site-packages/modelscope"} \ No newline at end of file