identifier
stringlengths
24
102
embedding
listlengths
2.56k
2.56k
tokens
listlengths
4
448
fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerRelPositionalEncoding
[ -0.0003179285558871925, 0.045319270342588425, 0.02566550485789776, 0.00661869440227747, -0.0014668067451566458, 0.027052829042077065, 0.009884688071906567, -0.002413366688415408, 0.0007406290387734771, 0.015954231843352318, 0.01907571218907833, 0.008034921251237392, 0.005780518986284733, 0...
[ "Dropout", "False", "ModelRelPositionalEncoding", "Module", "None", "__init__", "arange", "cat", "center_idx", "class", "config", "cos", "def", "device", "dim", "div_term", "dropout", "dtype", "embed_dim", "exp", "expand", "extend_pos_enc", "feature_representation", "fl...
fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerEncoder
[ -0.00019217580847907811, 0.037181150168180466, 0.025165289640426636, 0.027772504836320877, -0.0008678907761350274, 0.0446627214550972, 0.018590575084090233, -0.031059861183166504, 0, -0.010088787414133549, 0.03264686092734337, 0.010315501131117344, -0.0017216118285432458, 0.017457002773880...
[ "BaseModelOutput", "Embedding", "False", "ModelEncoder", "ModelEncoderLayer", "ModelRelPositionalEncoding", "Module", "ModuleList", "None", "_", "__init__", "all_hidden_states", "all_self_attentions", "attention_mask", "attentions", "class", "config", "conformer_layer", "conforme...
fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerLoss
[ -0.00010755642870208248, 0.030344083905220032, 0.004326841793954372, -0.009384188801050186, -0.00020369871344882995, 0.04675236716866493, 0.0206789318472147, -0.013261488638818264, 0.010114694945514202, -0.013261488638818264, 0.03776152804493904, 0.027534445747733116, -0.00014311590348370373...
[ "False", "L1Loss", "MSELoss", "ModelLoss", "Module", "None", "True", "__init__", "and", "both", "but", "class", "config", "def", "dim", "duration_criterion", "duration_labels", "duration_loss", "duration_mask", "duration_outputs", "duration_weights", "else", "energy_label...
fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerPreTrainedModel
[ -0.0001600405666977167, 0.045022401958703995, 0.00024094020773191005, 0.011593269184231758, -0.0006823707954026759, 0.03129056841135025, 0.01435089111328125, 0.003714348189532757, 0.0022792592644691467, 0.005121298134326935, 0.007428696379065514, -0.005627800244837999, -0.0009637608309276402...
[ "BatchNorm1d", "Conv1d", "Embedding", "False", "LayerNorm", "Linear", "Model", "ModelAttention", "ModelConfig", "ModelEncoder", "ModelPreTrainedModel", "ModelRelPositionalEncoding", "None", "PreTrainedModel", "_init_weights", "_is_hf_initialized", "_set_gradient_checkpointing", "a"...
fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerModel
[ 0.00001139327468990814, 0.05272457003593445, -0.016153911128640175, 0.0036738929338753223, 0.0000972810375969857, 0.03836553916335106, 0.047788653522729874, -0.011217993684113026, 0.007291696034371853, -0.03477577865123749, 0.05048097297549248, 0.00992792472243309, -0.0027624310459941626, ...
[ "Embedding", "False", "Linear", "ModelDurationPredictor", "ModelEncoder", "ModelLoss", "ModelModel", "ModelModelOutput", "ModelPreTrainedModel", "ModelSpeechDecoderPostnet", "ModelVarianceEmbedding", "ModelVariancePredictor", "None", "True", "__init__", "and", "any", "attention_mas...
fastspeech2_conformer/modeling_fastspeech2_conformer.py:HifiGanResidualBlock
[ -0.00004691328285844065, 0.012236400507390499, -0.018014701083302498, -0.024132901802659035, 0.0004000906483270228, 0.04645300284028053, 0.013029500842094421, -0.02152700163424015, 0.011613250710070133, -0.019034400582313538, 0.023226501420140266, -0.012123100459575653, 0.0012038125423714519...
[ "Conv1d", "ModelGanResidualBlock", "Module", "ModuleList", "_", "__init__", "apply_weight_norm", "channels", "class", "conv1", "conv2", "convs1", "convs2", "def", "dilation", "for", "forward", "functional", "get_padding", "hasattr", "hidden_states", "i", "if", "in", "...
fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerHifiGan
[ -0.00028376519912853837, 0.03287366032600403, -0.021379372105002403, 0.013908086344599724, -0.0011135089443996549, -0.010517272166907787, 0.058161091059446335, -0.02666674368083477, 0.0018247179687023163, -0.0317242294549942, 0.0349426306784153, 0.002025868045166135, -0.0035632287617772818, ...
[ "Conv1d", "ConvTranspose1d", "HifiGanResidualBlock", "ModelHifiGan", "ModelHifiGanConfig", "ModuleList", "PreTrainedModel", "__init__", "_init_weights", "append", "apply_weight_norm", "auto_docstring", "channels", "class", "config", "conv_post", "conv_pre", "custom_intro", "def",...
fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerWithHifiGan
[ -0.00019394901755731553, 0.03718153014779091, -0.029019732028245926, 0.008728590793907642, -0.0007757960702292621, 0.004024220630526543, 0.08161799609661102, -0.012696132995188236, 0.0034290894400328398, -0.013999753631651402, 0.052371546626091, 0.01983770728111267, -0.004165918566286564, ...
[ "ModelHifiGan", "ModelModel", "ModelWithHifiGan", "ModelWithHifiGanConfig", "ModelWithHifiGanOutput", "None", "PreTrainedModel", "__init__", "attention_mask", "auto_docstring", "class", "config", "def", "duration_labels", "else", "energy_labels", "forward", "has_missing_labels", ...
fast_vlm/modeling_fast_vlm.py:FastVlmMultiModalProjector
[ -0.00025449725217185915, 0.037359483540058136, 0.04487694054841995, 0.029386423528194427, -0.0008898505475372076, 0.03644827753305435, 0.027677910402417183, -0.03644827753305435, -0.0013952855952084064, 0.009795474819839, 0.03280344977974892, -0.009852425195276737, -0.0020786907989531755, ...
[ "ACT2FN", "Linear", "ModelMultiModalProjector", "Module", "__init__", "act", "class", "config", "def", "forward", "hidden_size", "hidden_states", "image_features", "linear_1", "linear_2", "nn", "projector_hidden_act", "return", "self", "super", "text_config", "vision_config...
fast_vlm/modeling_fast_vlm.py:FastVlmPreTrainedModel
[ -0.00040332425851374865, 0.03276687115430832, 0.01826929859817028, 0.019447963684797287, -0.0021068628411740065, 0.03158820793032646, 0.028523679822683334, -0.03323833644390106, -0.003683326533064246, 0.01744423434138298, 0.015440504997968674, 0.008898916654288769, -0.004272658843547106, 0...
[ "ModelConfig", "ModelPreTrainedModel", "PreTrainedModel", "True", "_can_compile_fullgraph", "_skip_keys_device_placement", "_supports_attention_backend", "_supports_flash_attn", "_supports_flex_attn", "_supports_sdpa", "base_model_prefix", "class", "config", "image", "input_modalities", ...
fast_vlm/modeling_fast_vlm.py:FastVlmModelOutputWithPast
[ -0.000123380494187586, 0.02538112923502922, 0.014777458272874355, 0.01703355833888054, -0.0007931602885946631, 0.02910369634628296, 0.04331712797284126, -0.038353707641363144, 0.016920752823352814, -0.007952754385769367, 0.02763723023235798, 0.026170765981078148, -0.002354804892092943, 0.0...
[ "BaseModelOutputWithPast", "ModelModelOutputWithPast", "None", "class", "image_hidden_states", "r" ]
fast_vlm/modeling_fast_vlm.py:FastVlmModel
[ -0.00011488077143440023, 0.03876703605055809, 0.0003655297332443297, 0.01815812475979328, -0.000342901679687202, 0.029186677187681198, 0.02005191519856453, -0.036761846393346786, -0.0013855317374691367, 0.02940947748720646, 0.030077874660491943, 0.010471556335687637, -0.0002436864742776379, ...
[ "AutoModel", "False", "Image", "ModelModel", "ModelModelOutputWithPast", "ModelMultiModalProjector", "ModelPreTrainedModel", "None", "Obtains", "True", "ValueError", "You", "__init__", "_checkpoint_conversion_mapping", "all", "and", "apply", "attention_mask", "attentions", "aut...
fast_vlm/modeling_fast_vlm.py:FastVlmCausalLMOutputWithPast
[ -0.00020158913685008883, 0.02637176588177681, 0.02978190779685974, 0.009434727020561695, -0.0011793408775702119, 0.03387407958507538, 0.03773890808224678, -0.030463937669992447, 0.01784641109406948, -0.004205842036753893, 0.020006166771054268, 0.018642110750079155, -0.0008596400148235261, ...
[ "ModelCausalLMOutputWithPast", "ModelOutput", "None", "attentions", "class", "hidden_states", "image_hidden_states", "logits", "loss", "past_key_values", "r" ]
fast_vlm/modeling_fast_vlm.py:FastVlmForConditionalGeneration
[ -0.00021033240773249418, 0.04055279120802879, 0.009180701337754726, 0.02230403572320938, -0.0009011731599457562, 0.023881088942289352, 0.024669615551829338, -0.04122867062687874, -0.00328083336353302, 0.018361402675509453, 0.04122867062687874, 0.003689177567139268, 0.002802085131406784, -0...
[ "False", "GenerationMixin", "Linear", "ModelCausalLMOutputWithPast", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "True", "__init__", "_checkpoint_conversion_mapping", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "cache_posit...
grounding_dino/modeling_grounding_dino.py:MultiScaleDeformableAttention
[ -0.00033676877501420677, 0.030128130689263344, 0.006605013273656368, 0.00004798049485543743, -0.0014919218374416232, 0.012746516615152359, 0.050290804356336594, -0.044728685170412064, 0.007705848664045334, 0.02688356302678585, 0.008632868528366089, 0.02097381465137005, -0.002129247644916177,...
[ "False", "ModelScaleDeformableAttention", "Module", "_", "align_corners", "append", "attention_weights", "batch_size", "bilinear", "class", "contiguous", "def", "dim", "enumerate", "flatten", "for", "forward", "functional", "grid_sample", "height", "hidden_dim", "im2col_ste...
grounding_dino/modeling_grounding_dino.py:GroundingDinoDecoderOutput
[ -0.0001767648645909503, 0.05502983182668686, 0.020693035796284676, 0.01341636385768652, -0.0011298739118501544, 0.03524637967348099, 0.06549004465341568, -0.04547920078039169, 0.01569032482802868, -0.010346517898142338, -0.00739037012681365, 0.028424499556422234, -0.0036667603999376297, 0....
[ "ModelDecoderOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "intermediate_hidden_states", "intermediate_reference_points", "last_hidden_state", "r" ]
grounding_dino/modeling_grounding_dino.py:GroundingDinoEncoderOutput
[ -0.00012691110896412283, 0.031812384724617004, 0.023238832131028175, 0.03429419919848442, -0.0005428974982351065, 0.04489832744002342, 0.05144130066037178, -0.04444708675146103, 0.020192965865135193, 0.013480779714882374, 0.022336354479193687, 0.05144130066037178, -0.001205655513331294, 0....
[ "ModelEncoderOutput", "ModelOutput", "None", "attentions", "class", "last_hidden_state_text", "last_hidden_state_vision", "r", "text_hidden_states", "vision_hidden_states" ]
grounding_dino/modeling_grounding_dino.py:GroundingDinoModelOutput
[ -0.00005340695133781992, 0.038869667798280716, 0.014067655429244041, 0.028474289923906326, -0.0005473108612932265, 0.05423674359917641, 0.05242885276675224, -0.0223726574331522, 0.017174968495965004, -0.002937823534011841, 0.0011652425164356828, 0.024858508259058, -0.002316360827535391, 0....
[ "ModelModelOutput", "ModelOutput", "None", "class", "decoder_attentions", "decoder_hidden_states", "enc_outputs_class", "enc_outputs_coord_logits", "encoder_attentions", "encoder_last_hidden_state_text", "encoder_last_hidden_state_vision", "encoder_logits", "encoder_pred_boxes", "encoder_t...
grounding_dino/modeling_grounding_dino.py:GroundingDinoObjectDetectionOutput
[ -0.00013554371253121644, 0.018483556807041168, 0.004224002826958895, 0.004450794775038958, -0.0007122689276002347, 0.05511048063635826, 0.0566980242729187, -0.020524684339761734, 0.014628089964389801, -0.009978852234780788, 0, 0.034472398459911346, -0.0044791437685489655, 0.007881025783717...
[ "ModelObjectDetectionOutput", "ModelOutput", "None", "auxiliary_outputs", "class", "decoder_attentions", "decoder_hidden_states", "enc_outputs_class", "enc_outputs_coord_logits", "encoder_attentions", "encoder_last_hidden_state_text", "encoder_last_hidden_state_vision", "encoder_logits", "...
grounding_dino/modeling_grounding_dino.py:GroundingDinoFrozenBatchNorm2d
[ -0.00022441272449214011, 0.023253431543707848, 0.018465960398316383, 0.0027356978971511126, -0.0008050361066125333, 0.029408752918243408, -0.01937786117196083, -0.014704376459121704, 0.01687013730406761, 0.020061785355210304, 0.01607222482562065, -0.003034914843738079, 0.00237948726862669, ...
[ "ModelFrozenBatchNorm2d", "Module", "__init__", "_load_from_state_dict", "bias", "class", "def", "del", "epsilon", "error_msgs", "forward", "if", "in", "local_metadata", "missing_keys", "n", "nn", "num_batches_tracked", "num_batches_tracked_key", "ones", "prefix", "register...
grounding_dino/modeling_grounding_dino.py:replace_batch_norm
[ 0.00006147739622974768, 0.005075003486126661, 0.02018596976995468, -0.013628380373120308, 0.0004401425248943269, 0.03170451521873474, -0.014312650077044964, -0.013457313179969788, 0.028169121593236923, 0.00610140897333622, 0.011119389906525612, -0.007698039058595896, 0.0019815322011709213, ...
[ "BatchNorm2d", "ModelFrozenBatchNorm2d", "Model_batch_norm", "_modules", "bias", "children", "copy_", "def", "device", "for", "if", "in", "isinstance", "len", "list", "meta", "model", "module", "name", "named_children", "new_module", "nn", "num_features", "r", "runnin...
grounding_dino/modeling_grounding_dino.py:GroundingDinoConvEncoder
[ -0.0001403618516633287, 0.02088584378361702, 0.03121647611260414, 0.011116658337414265, -0.0004982845857739449, 0.030767317861318588, -0.0028212731704115868, -0.03660637140274048, 0.0035090462770313025, 0.013250159099698067, 0.013587026856839657, 0.01235184259712696, -0.00013158922956790775,...
[ "ModelConvEncoder", "Module", "None", "True", "__init__", "and", "append", "backbone", "backbone_config", "backbone_kwargs", "backbone_model_type", "bool", "channels", "class", "config", "create_model", "def", "elif", "else", "feature_info", "feature_map", "feature_maps", ...
grounding_dino/modeling_grounding_dino.py:GroundingDinoConvModel
[ -0.00013853727432433516, 0.01874930039048195, 0.0037837442941963673, 0.010165282525122166, -0.0004941456718370318, 0.040209341794252396, 0.01999172382056713, -0.024735521525144577, 0.010391178540885448, 0.006183880381286144, 0.012367760762572289, 0.02789805456995964, 0.0018353982595726848, ...
[ "ModelConvModel", "Module", "__init__", "class", "conv_encoder", "def", "dtype", "feature_map", "for", "forward", "in", "mask", "nn", "out", "pixel_mask", "pixel_values", "pos", "position_embedding", "return", "self", "super", "to" ]
grounding_dino/modeling_grounding_dino.py:GroundingDinoSinePositionEmbedding
[ -0.00008343662193510681, 0.03642403706908226, 0.02653108909726143, 0.01944863796234131, -0.0005094025400467217, 0.006407932844012976, 0.05261249840259552, -0.012478605844080448, 0.01506426278501749, 0.01641330122947693, 0.0252944715321064, 0.026193829253315926, 0.0017706130165606737, -0.00...
[ "ModelSinePositionEmbedding", "Module", "None", "__init__", "arange", "cat", "class", "config", "cos", "cumsum", "d_model", "def", "device", "dim", "dim_t", "div", "dtype", "embedding_dim", "eps", "flatten", "float32", "floor", "forward", "math", "nn", "permute", ...
grounding_dino/modeling_grounding_dino.py:GroundingDinoLearnedPositionEmbedding
[ -0.0002529557386878878, -0.002436925657093525, 0.014707059599459171, 0.024283749982714653, -0.001083077979274094, 0.030098168179392815, 0.036938659846782684, -0.013623981736600399, 0.008835637010633945, 0.017443256452679634, 0.0010474504670128226, 0.02462577447295189, 0.0016459935577586293, ...
[ "Embedding", "ModelLearnedPositionEmbedding", "Module", "None", "__init__", "arange", "cat", "class", "column_embeddings", "config", "d_model", "def", "device", "dim", "embedding_dim", "forward", "height", "height_values", "nn", "permute", "pixel_mask", "pixel_values", "p...
grounding_dino/modeling_grounding_dino.py:build_position_encoding
[ -0.0001506119588157162, 0.05156097933650017, 0.004819582682102919, 0.02794787473976612, -0.0008270289399661124, 0.02144571579992771, 0.048366934061050415, 0.002823305781930685, 0.010836930945515633, 0.011692478321492672, 0.015742069110274315, 0.004505882039666176, 0.004135144874453545, -0....
[ "ModelLearnedPositionEmbedding", "ModelSinePositionEmbedding", "Model_position_encoding", "config", "def", "elif", "else", "if", "learned", "position_embedding", "position_embedding_type", "return", "sine" ]
grounding_dino/modeling_grounding_dino.py:GroundingDinoMultiscaleDeformableAttention
[ -0.00041481046355329454, 0.05053410306572914, 0.023054728284478188, -0.007976004853844643, -0.0015573585405945778, 0.019328711554408073, 0.023404043167829514, -0.03912317752838135, -0.0004912228905595839, 0.024219108745455742, -0.011178050190210342, 0.02759581245481968, -0.001557358540594577...
[ "CUDA", "F", "False", "Linear", "Make", "ModelMultiscaleDeformableAttention", "Module", "MultiScaleDeformableAttention", "None", "ValueError", "You", "_", "__init__", "a", "align", "and", "attention", "attention_mask", "attention_weights", "attn", "authors", "batch_size", ...
grounding_dino/modeling_grounding_dino.py:GroundingDinoTextEnhancerLayer
[ -0.0001596019574208185, 0.05747073516249657, 0.03704168647527695, -0.004153158515691757, -0.0004875752201769501, 0.04287855699658394, 0.016388140618801117, -0.007464461028575897, 0.003199054626747966, 0.008137946017086506, 0.018184099346399307, 0.03165380284190178, 0.0037883541081100702, -...
[ "ACT2FN", "LayerNorm", "Linear", "ModelMultiheadAttention", "ModelTextEnhancerLayer", "Module", "None", "True", "__init__", "activation", "activation_function", "and", "attention_mask", "attention_masks", "attention_output", "attention_weights", "class", "config", "d_model", "d...
grounding_dino/modeling_grounding_dino.py:GroundingDinoBiMultiHeadAttention
[ -0.00011811069998657331, 0.04658636078238487, 0.05285760015249252, -0.008063023909926414, -0.00020210053480695933, 0.018253790214657784, 0.017917830497026443, -0.03695552423596382, 0.0033315967302769423, 0.03135620430111885, 0.03292401507496834, 0.02766065113246441, -0.00038320361636579037, ...
[ "Attention", "F", "Linear", "ModelBiMultiHeadAttention", "Module", "None", "True", "ValueError", "_", "__init__", "_reshape", "and", "attn_weights", "attn_weights_transposed", "batch_size", "be", "bmm", "but", "by", "clamp", "class", "config", "contiguous", "d_model", ...
grounding_dino/modeling_grounding_dino.py:drop_path
[ 0, 0.017753854393959045, 0.03775503486394882, -0.02123720571398735, 0, 0.043598074465990067, 0.0339345820248127, -0.027304979041218758, 0.014832334592938423, 0.0017486985307186842, 0.020113544538617134, -0.04944111406803131, 0.0016574009787291288, -0.018315685912966728, -0.05034004524350...
[ "False", "Model_path", "Model_prob", "def", "device", "div", "dtype", "floor_", "if", "input", "keep_prob", "ndim", "not", "or", "output", "rand", "random_tensor", "return", "shape", "torch", "training" ]
grounding_dino/modeling_grounding_dino.py:GroundingDinoDropPath
[ -0.00011107319733127952, 0.018094833940267563, 0.03596488758921623, -0.025063030421733856, -0.00047414645086973906, 0.031918834894895554, 0.033717080950737, -0.026861274614930153, 0.0066872211173176765, -0.002430439693853259, 0.009721758775413036, -0.033717080950737, -0.000818341679405421, ...
[ "ModelDropPath", "Module", "None", "__init__", "class", "def", "drop_path", "drop_prob", "extra_repr", "f", "forward", "hidden_states", "nn", "p", "return", "self", "super", "training" ]
grounding_dino/modeling_grounding_dino.py:GroundingDinoFusionLayer
[ -0.00009035070979734883, 0.05524200201034546, 0.03615431860089302, 0.01751575618982315, -0.00009649105049902573, 0.02986661158502102, 0.0294174887239933, -0.04378939047455788, 0.007242091931402683, 0.02211925759911537, 0.022456098347902298, 0.018975403159856796, 0.003185959067195654, -0.00...
[ "Identity", "LayerNorm", "ModelBiMultiHeadAttention", "ModelDropPath", "ModelFusionLayer", "Module", "None", "Parameter", "True", "__init__", "attention_mask_text", "attention_mask_vision", "attn", "class", "config", "d_model", "def", "delta_t", "delta_v", "drop_path", "else"...
grounding_dino/modeling_grounding_dino.py:GroundingDinoDeformableLayer
[ -0.00034602233790792525, 0.06585656106472015, 0.02040625922381878, -0.0015797459054738283, -0.001427568611688912, 0.026435380801558495, 0.020870037376880646, -0.027362938970327377, 0.004550827667117119, 0.013681469485163689, 0.0047827172093093395, 0.022261373698711395, 0.00043298082891851664...
[ "ACT2FN", "False", "LayerNorm", "Linear", "ModelDeformableLayer", "ModelMultiscaleDeformableAttention", "Module", "None", "__init__", "activation_dropout", "activation_fn", "activation_function", "any", "attention_mask", "attn_weights", "clamp", "clamp_value", "class", "config", ...
grounding_dino/modeling_grounding_dino.py:get_sine_pos_embed
[ -0.00020325295918155462, 0.04230489209294319, 0.01674097403883934, 0.019681954756379128, -0.0007670589839108288, 0.007069667801260948, 0.051354069262742996, 0.00882294587790966, 0.011763927526772022, 0.007352454587817192, -0.004298358224332333, -0.004468030296266079, -0.000043743570131482556...
[ "Model_sine_pos_embed", "True", "arange", "cat", "cos", "def", "device", "dim", "dim_t", "div", "dtype", "exchange_xy", "flatten", "float32", "floor", "for", "if", "in", "math", "num_pos_feats", "pi", "pos_tensor", "position_embeddings", "return", "rounding_mode", "...
grounding_dino/modeling_grounding_dino.py:GroundingDinoEncoderLayer
[ -0.0001626005832804367, 0.049769919365644455, -0.0012371783377602696, 0.019568627700209618, -0.00049487134674564, 0.03755366429686546, 0.026694774627685547, -0.03189799189567566, 0.005118383560329676, 0.020812874659895897, 0.012385922484099865, 0.0411732941865921, 0.001081647351384163, 0.0...
[ "False", "ModelDeformableLayer", "ModelEncoderLayer", "ModelFusionLayer", "ModelTextEnhancerLayer", "Module", "None", "_", "__init__", "and", "arange", "attention_mask", "attention_mask_text", "attention_mask_vision", "attention_masks", "batch_size", "class", "config", "d_model",...
grounding_dino/modeling_grounding_dino.py:GroundingDinoMultiheadAttention
[ -0.00019792768580373377, 0.022054798901081085, 0.047728847712278366, 0.0006397305405698717, -0.0007987795397639275, 0.01583421416580677, 0.03732350468635559, -0.01662592589855194, 0.0013289429480209947, 0.025447843596339226, 0.0066164396703243256, 0.005683351773768663, -0.003378908149898052,...
[ "Dropout", "False", "Linear", "ModelMultiheadAttention", "Module", "None", "The", "ValueError", "_", "__init__", "a", "all_head_size", "and", "attention", "attention_dropout", "attention_head_size", "attention_mask", "attention_probs", "attention_scores", "batch_size", "class...
grounding_dino/modeling_grounding_dino.py:GroundingDinoDecoderLayer
[ -0.00014667253708466887, 0.06469142436981201, 0.016738340258598328, 0.0026153656654059887, -0.0006750471075065434, 0.032571908086538315, 0.044107791036367416, -0.045238759368658066, 0.007973331026732922, 0.0019226472359150648, 0.0057396674528717995, 0.030988549813628197, -0.00156215089373290...
[ "ACT2FN", "False", "LayerNorm", "Linear", "ModelDecoderLayer", "ModelMultiheadAttention", "ModelMultiscaleDeformableAttention", "Module", "None", "True", "__init__", "activation_dropout", "activation_fn", "activation_function", "attention_mask", "class", "config", "cross_attn_weigh...
grounding_dino/modeling_grounding_dino.py:GroundingDinoContrastiveEmbedding
[ -0.000055858748964965343, 0.04915569722652435, 0.0790959894657135, 0.0028627607971429825, -0.00038228329503908753, 0.025024719536304474, 0.016757624223828316, -0.03217463940382004, 0.015752166509628296, 0.03329181298613548, 0.038654252886772156, 0.02212006412446499, 0.00135457469150424, -0...
[ "ModelContrastiveEmbedding", "Module", "None", "__init__", "class", "config", "def", "device", "float", "forward", "full", "inf", "masked_fill", "max_text_len", "new_output", "nn", "output", "return", "self", "shape", "super", "text_hidden_state", "text_token_mask", "to...
grounding_dino/modeling_grounding_dino.py:GroundingDinoPreTrainedModel
[ -0.0002789895806927234, 0.05905994772911072, 0.0006187845719978213, 0.0135631849989295, -0.0012375691439956427, 0.028957687318325043, 0.028614314272999763, -0.03342152014374733, 0.003505253465846181, 0.007611407898366451, 0.00589454872533679, -0.00184562336653471, -0.0031761890277266502, -...
[ "Conv2d", "Embedding", "False", "GroupNorm", "LayerNorm", "Linear", "ModelBiMultiHeadAttention", "ModelConfig", "ModelDecoder", "ModelFusionLayer", "ModelLearnedPositionEmbedding", "ModelMLPPredictionHead", "ModelMultiscaleDeformableAttention", "ModelPreTrainedModel", "None", "PreTrain...
grounding_dino/modeling_grounding_dino.py:GroundingDinoEncoder
[ -0.000011243238986935467, 0.047849468886852264, 0.011513076722621918, 0.029653193429112434, 0, 0.04829876124858856, 0.02662047930061817, -0.04650159925222397, 0.010614494793117046, 0.022015249356627464, 0.016623759642243385, 0.03526932746171951, 0.0008985815802589059, 0.002232413738965988,...
[ "ModelEncoder", "ModelEncoderLayer", "ModelEncoderOutput", "ModelPreTrainedModel", "ModuleList", "None", "_", "__init__", "all_attn_deformable", "all_attn_enhanced_text", "all_attn_fused_text", "all_attn_fused_vision", "all_attns", "append", "attentions", "cat", "class", "config", ...
grounding_dino/modeling_grounding_dino.py:GroundingDinoDecoder
[ -0.00006920980376889929, 0.05965710058808327, 0.0014718034071847796, 0.007737481035292149, -0.0003942330658901483, 0.030277099460363388, 0.05046183243393898, -0.04440641403198242, 0.009755954146385193, -0.004934045951813459, -0.0021025764290243387, 0.02197893150150776, -0.0038687405176460743...
[ "False", "Last", "LayerNorm", "ModelDecoder", "ModelDecoderLayer", "ModelDecoderOutput", "ModelMLPPredictionHead", "ModelPreTrainedModel", "ModuleList", "None", "ValueError", "_", "__init__", "all_attns", "all_cross_attns_text", "all_cross_attns_vision", "all_hidden_states", "all_s...
grounding_dino/modeling_grounding_dino.py:generate_masks_with_special_tokens_and_transfer_map
[ -0.000059773941757157445, 0.009918046183884144, -0.017002366483211517, -0.07753078639507294, -0.00040557727334089577, 0.042619261890649796, 0.00535574508830905, -0.046926528215408325, 0.006942632608115673, 0.0055541060864925385, 0.03309793770313263, 0.03944548964500427, 0.001197249861434102,...
[ "Model_masks_with_special_tokens_and_transfer_map", "SPECIAL_TOKENS", "arange", "attention_mask", "batch_size", "bool", "clamp", "cummax", "cummin", "def", "device", "dim", "dims", "dtype", "expand", "eye", "flip", "identity", "indices", "input_ids", "isin", "long", "min"...
grounding_dino/modeling_grounding_dino.py:GroundingDinoModel
[ -0.00005853917173226364, 0.036507315933704376, 0.004957783501595259, 0.02005648799240589, -0.00039789031143300235, 0.053408849984407425, 0.02287341095507145, -0.013690243475139141, 0.013352212496101856, 0.003464814741164446, 0.014873350970447063, 0.018479011952877045, -0.0006056383135728538,...
[ "AutoModel", "Conv2d", "Embedding", "False", "GroupNorm", "LayerNorm", "Linear", "ModelContrastiveEmbedding", "ModelConvEncoder", "ModelConvModel", "ModelDecoder", "ModelEncoder", "ModelEncoderOutput", "ModelMLPPredictionHead", "ModelModel", "ModelModelOutput", "ModelPreTrainedModel"...
grounding_dino/modeling_grounding_dino.py:GroundingDinoMLPPredictionHead
[ -0.00024052357184700668, -0.0006734660128131509, 0.02770831435918808, 0.04241766780614853, -0.0006734660128131509, 0.040137145668268204, 0.04583844542503357, -0.009635196067392826, -0.0012685391120612621, -0.013569092378020287, 0.022007014602422714, -0.0588374100625515, -0.002750876825302839...
[ "Linear", "ModelMLPPredictionHead", "Module", "ModuleList", "__init__", "class", "def", "else", "enumerate", "for", "forward", "functional", "h", "hidden_dim", "i", "if", "in", "input_dim", "k", "layer", "layers", "n", "nn", "num_layers", "output_dim", "relu", "re...
grounding_dino/modeling_grounding_dino.py:build_label_maps
[ -0.00016861100448295474, -0.002711900509893894, -0.016610391438007355, -0.015367437154054642, -0.00041314109694212675, 0.0470062755048275, 0.012034058570861816, -0.03548070043325424, 0.0050848135724663734, -0.03412475064396858, 0.06327768415212631, 0.021243222057819366, -0.000169493781868368...
[ "F", "Model_label_maps", "None", "SPECIAL_TOKENS", "cumsum", "def", "delimiter_token_masks", "delimiter_tokens", "device", "dim", "for", "in", "input_ids", "int32", "isin", "label_group", "label_groups", "label_map", "label_maps", "logits", "max_seq_len", "num_labels", "p...
grounding_dino/modeling_grounding_dino.py:build_text_mask
[ 0.00013015992590226233, 0.029493534937500954, 0.03827405348420143, -0.023414714261889458, 0.0006472817622125149, 0.03647292032837868, 0.0059662493877112865, -0.04615400359034538, 0.0167730413377285, -0.009343371726572514, 0.04637914523482323, 0.02409013919532299, 0.0007633703644387424, -0....
[ "Model_text_mask", "None", "attention_mask", "bool", "def", "device", "dtype", "logits", "return", "seq_len", "shape", "text_mask", "torch", "zeros_like" ]
grounding_dino/modeling_grounding_dino.py:GroundingDinoForObjectDetection
[ -0.00012111647083656862, 0.0260628592222929, 0.004353171680122614, -0.0011023354018107057, -0.0004124981351196766, 0.03482537344098091, 0.03527473285794258, -0.023703722283244133, 0.009998252615332603, -0.0065999701619148254, 0.018648425117135048, 0.048306163400411606, -0.0012076541315764189...
[ "FloatTensor", "LongTensor", "ModelContrastiveEmbedding", "ModelForObjectDetection", "ModelMLPPredictionHead", "ModelModel", "ModelObjectDetectionOutput", "ModelPreTrainedModel", "ModuleList", "None", "_", "__init__", "_tied_weights_keys", "append", "attention_mask", "auto_docstring", ...
speecht5/modeling_speecht5.py:shift_tokens_right
[ -0.0001357423752779141, 0.026739485561847687, 0.00459761219099164, -0.04174518957734108, -0.0005782272783108056, 0.05189942196011543, 0.04039129242300987, -0.057089366018772125, 0.006572046782821417, 0.007784913759678602, 0.020759768784046173, -0.009872172959148884, -0.0006698974757455289, ...
[ "Model_tokens_right", "Modeled_input_ids", "None", "clone", "decoder_start_token_id", "def", "if", "input_ids", "is", "masked_fill_", "new_zeros", "pad_token_id", "return", "shape" ]
speecht5/modeling_speecht5.py:shift_spectrograms_right
[ -0.00009768103336682543, 0.03892390429973602, 0.009787551127374172, -0.009052070789039135, -0.0003147008828818798, 0.03915020450949669, 0.02206442318856716, -0.04571295902132988, 0.01035330630838871, 0.005827270448207855, 0.020480310544371605, 0.017990991473197937, 0.00047735529369674623, ...
[ "Model_spectrograms_right", "Modeled_input_values", "None", "attention_mask", "clone", "def", "if", "input_values", "is", "masked_fill_", "new_zeros", "not", "reduction_factor", "return", "shape" ]
speecht5/modeling_speecht5.py:_compute_mask_indices
[ 0, -0.008906038478016853, 0.00551726296544075, -0.014171244576573372, 0, 0.024309566244482994, -0.0006616514874622226, -0.05265205353498459, 0.013331051915884018, -0.0200525913387537, 0.040777336806058884, 0.030695026740431786, -0.006189417093992233, -0.012434846721589565, 0.005489256698...
[ "False", "None", "ValueError", "_", "_compute_mask_indices", "and", "append", "arange", "array", "attention_mask", "batch_size", "be", "bigger", "bool", "broadcast_to", "but", "choice", "compute_num_masked_span", "concatenate", "def", "detach", "dtype", "dummy_mask_idx", ...
speecht5/modeling_speecht5.py:SpeechT5NoLayerNormConvLayer
[ 0.000045173001126386225, 0.0243696216493845, 0.017374638468027115, 0.00046539207687601447, 0.00048302055802196264, 0.023467043414711952, 0.009871953167021275, -0.03474927693605423, 0.016584880650043488, 0.0016359236324205995, 0.02876969240605831, 0.00823602918535471, 0.004118014592677355, ...
[ "ACT2FN", "Conv1d", "GradientCheckpointingLayer", "ModelNoLayerNormConvLayer", "__init__", "activation", "class", "config", "conv", "conv_dim", "conv_kernel", "conv_stride", "def", "else", "feat_extract_activation", "forward", "hidden_states", "if", "in_conv_dim", "kernel_size"...
speecht5/modeling_speecht5.py:SpeechT5LayerNormConvLayer
[ -0.000011282235391263384, 0.01938651315867901, 0.025022126734256744, 0.010425886139273643, 0.000192843668628484, 0.03246113657951355, 0.011553009040653706, -0.028741631656885147, 0.016906842589378357, 0.009298763237893581, 0.030883165076375008, 0.015216158702969551, 0.00300096464343369, -0...
[ "ACT2FN", "Conv1d", "GradientCheckpointingLayer", "LayerNorm", "ModelLayerNormConvLayer", "True", "__init__", "activation", "class", "config", "conv", "conv_dim", "conv_kernel", "conv_stride", "def", "elementwise_affine", "else", "feat_extract_activation", "forward", "hidden_st...
speecht5/modeling_speecht5.py:SpeechT5GroupNormConvLayer
[ -0.000032619624107610434, 0.008350623771548271, -0.005811582785099745, 0.008237777277827263, 0.00018690162687562406, 0.031371261924505234, 0.009761202149093151, -0.023472024127840996, 0.013146590441465378, 0.0025390409864485264, 0.03746496140956879, 0.00668614124879241, 0.001227203174494207,...
[ "ACT2FN", "Conv1d", "GradientCheckpointingLayer", "GroupNorm", "ModelGroupNormConvLayer", "True", "__init__", "activation", "affine", "class", "config", "conv", "conv_dim", "conv_kernel", "conv_stride", "def", "else", "feat_extract_activation", "forward", "hidden_states", "if...
speecht5/modeling_speecht5.py:SpeechT5SinusoidalPositionalEmbedding
[ -0.00022105126117821783, 0.03214262053370476, 0.018334876745939255, 0.006083330605179071, -0.0010468987748026848, -0.00019540931680239737, 0.04549765586853027, -0.0075829424895346165, 0.0063096871599555016, -0.008771314285695553, 0.027615493163466454, 0.022522471845149994, -0.000636627664789...
[ "False", "ModelSinusoidalPositionalEmbedding", "Module", "None", "__init__", "arange", "bsz", "cat", "class", "cos", "create_position_ids_from_input_ids", "cumsum", "def", "detach", "device", "dim", "dtype", "emb", "emb_weights", "embedding_dim", "exp", "float", "forward"...
speecht5/modeling_speecht5.py:SpeechT5PositionalConvEmbedding
[ -0.00020198767015244812, 0.019476238638162613, 0.01247162651270628, -0.006064968649297953, -0.0009182875510305166, 0.020045718178153038, 0.021640263497829437, -0.03029637038707733, 0.012414677999913692, -0.00808662548661232, 0.020956886932253838, 0.014692600816488266, 0.0030040103010833263, ...
[ "ACT2FN", "Conv1d", "GatheredParameters", "ModelPositionalConvEmbedding", "ModelSamePadLayer", "Module", "__init__", "activation", "class", "config", "conv", "deepspeed", "def", "dim", "else", "feat_extract_activation", "forward", "groups", "hasattr", "hidden_size", "hidden_s...
speecht5/modeling_speecht5.py:SpeechT5ScaledPositionalEncoding
[ -0.0003658245550468564, 0.041933324187994, 0.03377962112426758, 0.011182219721376896, -0.0015579394530504942, 0.025975365191698074, 0.04868924990296364, 0.004251573234796524, 0.004979582037776709, 0.008037220686674118, 0.014268978498876095, 0.004921341314911842, 0.00030940386932343245, -0....
[ "Dropout", "False", "ModelScaledPositionalEncoding", "Module", "Parameter", "__init__", "alpha", "arange", "class", "cos", "def", "dim", "div_term", "dropout", "dtype", "emb", "exp", "float", "forward", "int64", "log", "math", "max_len", "nn", "p", "pe", "persiste...
speecht5/modeling_speecht5.py:SpeechT5RelativePositionalEncoding
[ -0.00024354932247661054, 0.02234923280775547, 0.052033599466085434, -0.007850884459912777, -0.0011389512801542878, 0.030257422477006912, 0.027277525514364243, -0.01008580718189478, 0.009111610241234303, 0.027621358633041382, 0.01971316896378994, 0.003954094834625721, 0.0034526698291301727, ...
[ "Embedding", "ModelRelativePositionalEncoding", "Module", "None", "__init__", "arange", "class", "def", "device", "dim", "dtype", "forward", "hidden_states", "long", "max_length", "nn", "pe_k", "pos_seq", "return", "self", "seq_len", "shape", "super", "to", "torch", ...
speecht5/modeling_speecht5.py:SpeechT5SamePadLayer
[ -0.00006862095324322581, 0.02618260867893696, 0.03647662326693535, -0.001608440070413053, -0.0004895252641290426, 0.007440784014761448, 0.009734558872878551, -0.04184741526842117, 0.01812642067670822, 0.0049232253804802895, -0.008112132549285889, -0.0022797889541834593, 0.002447626320645213,...
[ "ModelSamePadLayer", "Module", "__init__", "class", "def", "else", "forward", "hidden_states", "if", "nn", "num_conv_pos_embeddings", "num_pad_remove", "return", "self", "super" ]
speecht5/modeling_speecht5.py:SpeechT5FeatureEncoder
[ -0.0001271141809411347, 0.041128501296043396, -0.018643412739038467, 0.03186328709125519, -0.00009930795931722969, 0.029151519760489464, -0.001066346769221127, -0.03886869177222252, 0.007683346047997475, 0.008248298428952694, 0.021694153547286987, 0.007796336431056261, 0.0032908450812101364,...
[ "False", "ModelFeatureEncoder", "ModelGroupNormConvLayer", "ModelLayerNormConvLayer", "ModelNoLayerNormConvLayer", "Module", "ModuleList", "None", "True", "ValueError", "__init__", "_freeze_parameters", "_requires_grad", "and", "be", "but", "class", "config", "conv_layer", "con...
speecht5/modeling_speecht5.py:SpeechT5FeatureProjection
[ -0.00018812100461218506, 0.04205431789159775, 0.032105982303619385, 0.02882755547761917, -0.0005016559734940529, 0.021027158945798874, 0.026227423921227455, -0.03775844722986221, 0.0018935747211799026, 0.02532302960753441, 0.01820092648267746, 0.018992271274328232, 0.0018935747211799026, -...
[ "Dropout", "LayerNorm", "Linear", "ModelFeatureProjection", "Module", "__init__", "class", "config", "conv_dim", "def", "dropout", "eps", "feat_proj_dropout", "forward", "hidden_size", "hidden_states", "layer_norm", "layer_norm_eps", "nn", "norm_hidden_states", "projection", ...
speecht5/modeling_speecht5.py:SpeechT5SpeechEncoderPrenet
[ -0.00006349718023557216, 0.04416951537132263, 0.005156846717000008, 0.015806855633854866, -0.00021720414224546403, 0.018609490245580673, 0.016367383301258087, -0.03834003582596779, 0.0015274356119334698, -0.007118690758943558, 0.024326864629983902, 0.0345284529030323, 0.0004834543797187507, ...
[ "ModelFeatureEncoder", "ModelFeatureProjection", "ModelModelEncoderPrenet", "ModelPositionalConvEmbedding", "ModelSinusoidalPositionalEmbedding", "Module", "None", "Parameter", "Tensor", "True", "__init__", "_compute_mask_indices", "_conv_out_length", "_freeze_parameters", "_get_feat_ext...
speecht5/modeling_speecht5.py:SpeechT5SpeechDecoderPrenet
[ -0.00021967482462059706, 0.05989443138241768, -0.014402096159756184, 0.020803026854991913, -0.0009358504903502762, 0.015316515229642391, 0.051436055451631546, -0.03269047290086746, 0.001714535290375352, -0.004829274490475655, 0.036348145455121994, 0.01497360784560442, -0.0012073186226189137,...
[ "Linear", "ModelModelDecoderPrenet", "ModelScaledPositionalEncoding", "Model_decoder_prenet_dropout", "Model_decoder_prenet_layers", "Model_decoder_prenet_units", "Module", "ModuleList", "None", "__init__", "_consistent_dropout", "all_masks", "bernoulli", "cat", "class", "config", "d...
speecht5/modeling_speecht5.py:SpeechT5BatchNormConvLayer
[ -0.00013305226457305253, 0.012293676845729351, 0.02312113530933857, -0.02210606075823307, -0.0001885641395347193, 0.027745362371206284, 0.012068104930222034, -0.02312113530933857, 0.00981238391250372, -0.025376856327056885, 0.04894913360476494, -0.008346165530383587, 0.0021852292120456696, ...
[ "BatchNorm1d", "Conv1d", "Dropout", "ModelBatchNormConvLayer", "Model_decoder_postnet_dropout", "Model_decoder_postnet_kernel", "Model_decoder_postnet_layers", "Model_decoder_postnet_units", "Module", "None", "Tanh", "__init__", "activation", "batch_norm", "class", "config", "conv", ...
speecht5/modeling_speecht5.py:SpeechT5SpeechDecoderPostnet
[ -0.00021979870507493615, 0.034260235726833344, 0.008791947737336159, 0.007373891770839691, -0.0008862850954756141, 0.031083790585398674, 0.03539467975497246, -0.03312579169869423, 0.0011557157849892974, -0.024844342842698097, 0.03335268050432205, 0.017243562266230583, 0.00012939762382302433,...
[ "Linear", "ModelBatchNormConvLayer", "ModelModelDecoderPostnet", "Model_decoder_postnet_layers", "Module", "ModuleList", "__init__", "class", "config", "def", "feat_out", "for", "forward", "hidden_size", "hidden_states", "i", "in", "layer", "layer_output", "layers", "logits",...
speecht5/modeling_speecht5.py:SpeechT5TextEncoderPrenet
[ -0.0001684790477156639, 0.03405050188302994, 0.007008728571236134, 0.003617865964770317, -0.0010498905321583152, 0.023948853835463524, 0.04585467651486397, -0.021224813535809517, 0.004738694988191128, 0.005022449418902397, 0.025878382846713066, 0.0390445776283741, -0.0021565319038927555, -...
[ "Embedding", "EmbeddingAccessMixin", "ModelScaledPositionalEncoding", "ModelTextEncoderPrenet", "Module", "__init__", "class", "config", "def", "embed_tokens", "encode_positions", "forward", "hidden_size", "input_ids", "inputs_embeds", "max_text_positions", "nn", "pad_token_id", ...
speecht5/modeling_speecht5.py:SpeechT5TextDecoderPrenet
[ -0.00023370186681859195, 0.051150377839803696, 0.008106421679258347, -0.00030506119946949184, -0.0012559244642034173, 0.010047395713627338, 0.044984932988882065, -0.04087463393807411, 0.002212139777839184, -0.012844682671129704, 0.031968988478183746, 0.03402413800358772, -0.00142005097586661...
[ "Dropout", "Embedding", "EmbeddingAccessMixin", "ModelSinusoidalPositionalEmbedding", "ModelTextDecoderPrenet", "Module", "None", "__init__", "attention_mask", "class", "config", "def", "dropout", "else", "embed_positions", "embed_scale", "embed_tokens", "forward", "get_seq_lengt...
speecht5/modeling_speecht5.py:SpeechT5TextDecoderPostnet
[ -0.00029066388378851116, 0.04042021930217743, 0.025032976642251015, 0.02399950660765171, -0.0014784384984523058, 0.020439770072698593, 0.050525274127721786, -0.04524308815598488, -0.00044855536543764174, -0.006459197029471397, 0.014927921816706657, 0.0192914679646492, -0.0012703087413683534,...
[ "EmbeddingAccessMixin", "Linear", "ModelTextDecoderPostnet", "Module", "__init__", "class", "config", "def", "forward", "get_output_embeddings", "hidden_size", "hidden_states", "lm_head", "new_embeddings", "nn", "return", "self", "set_output_embeddings", "super", "vocab_size" ]
speecht5/modeling_speecht5.py:SpeechT5Attention
[ -0.0001474744494771585, 0.04247264191508293, 0.034607335925102234, -0.020000344142317772, -0.0006320333341136575, 0.023595912382006645, 0.025955503806471825, -0.01786547712981701, -0.0009129371028393507, 0.01696658506989479, 0.0026826304383575916, 0.016067693009972572, -0.002022506669163704,...
[ "Attention", "EncoderDecoderCache", "False", "Linear", "ModelAttention", "Module", "None", "True", "ValueError", "_", "__init__", "and", "attention_mask", "attn_output", "attn_probs", "attn_weights", "attn_weights_reshaped", "be", "bmm", "bsz", "but", "by", "cache_positio...
speecht5/modeling_speecht5.py:SpeechT5FeedForward
[ -0.00021927375928498805, 0.042442843317985535, 0.042442843317985535, 0.005989917553961277, -0.0009840579004958272, 0.025100605562329292, 0.039020031690597534, -0.02806704118847847, 0.0008806604892015457, -0.017342237755656242, 0.02943616546690464, -0.009355680085718632, 0.0024244904052466154...
[ "ACT2FN", "Dropout", "Linear", "ModelFeedForward", "Module", "__init__", "activation_dropout", "class", "config", "def", "else", "forward", "hidden_act", "hidden_dropout", "hidden_size", "hidden_states", "if", "intermediate_act_fn", "intermediate_dense", "intermediate_dropout",...
speecht5/modeling_speecht5.py:SpeechT5EncoderLayer
[ -0.00003288807056378573, 0.0347999632358551, 0.041086405515670776, 0.009485796093940735, -0.0003736084618140012, 0.03547351062297821, 0.0258193276822567, -0.011450310237705708, 0.005949670914560556, 0.0031151578295975924, 0.006567089818418026, 0.029860612004995346, 0.001894352724775672, -0...
[ "Dropout", "False", "GradientCheckpointingLayer", "LayerNorm", "ModelAttention", "ModelEncoderLayer", "ModelFeedForward", "None", "__init__", "attention", "attention_dropout", "attention_mask", "attn_weights", "class", "config", "def", "dropout", "embed_dim", "encoder_attention_h...
speecht5/modeling_speecht5.py:SpeechT5DecoderLayer
[ -0.00010298725828761235, 0.0522787980735302, 0.02546337991952896, -0.007830552756786346, -0.0006830607890151441, 0.027716776356101036, 0.034927643835544586, -0.029294153675436974, 0.0052954815328121185, -0.007154533639550209, -0.0045631276443600655, 0.02535071037709713, -0.000290476891677826...
[ "Dropout", "False", "GradientCheckpointingLayer", "LayerNorm", "ModelAttention", "ModelDecoderLayer", "ModelFeedForward", "None", "True", "__init__", "attention_dropout", "attention_mask", "cache_position", "class", "config", "cross_attn_weights", "decoder_attention_heads", "decode...
speecht5/modeling_speecht5.py:SpeechT5PreTrainedModel
[ -0.0002048766182269901, 0.05290055647492409, -0.019329048693180084, 0.009947112761437893, -0.0007771181990392506, 0.01729441247880459, 0.03187597543001175, -0.011190501973032951, 0, -0.016277093440294266, 0.019555119797587395, 0.0058778394013643265, -0.003871461609378457, -0.01661619916558...
[ "BatchNorm1d", "Conv1d", "Embedding", "False", "GroupNorm", "LayerNorm", "Linear", "Model", "ModelConfig", "ModelFeatureProjection", "ModelHifiGan", "ModelPositionalConvEmbedding", "ModelPreTrainedModel", "ModelScaledPositionalEncoding", "ModelSinusoidalPositionalEmbedding", "None", ...
speecht5/modeling_speecht5.py:SpeechT5Encoder
[ -0.00004495609027799219, 0.041094690561294556, 0.02705962024629116, 0.015831561759114265, -0.0002877189835999161, 0.04087013006210327, 0.009599989280104637, -0.010329813696444035, 0.0036210487596690655, -0.003508768044412136, 0.012463144026696682, 0.03166312351822853, -0.0003491224197205156,...
[ "BaseModelOutput", "Dropout", "False", "LayerNorm", "ModelEncoder", "ModelEncoderLayer", "ModelPreTrainedModel", "ModelRelativePositionalEncoding", "ModuleList", "None", "_", "__init__", "_prepare_4d_attention_mask", "all_hidden_states", "all_self_attentions", "attention_mask", "atte...
speecht5/modeling_speecht5.py:SpeechT5EncoderWithSpeechPrenet
[ 0.00007835146243451163, 0.034961383789777756, 0.009860903024673462, 0.02106647565960884, 0, 0.030479155480861664, 0.027901872992515564, -0.006583273410797119, 0.005182576831430197, 0.007619788870215416, 0.017368637025356293, 0.040115948766469955, 0.00026788320974446833, 0.00512654893100261...
[ "ModelEncoder", "ModelEncoderWithModelPrenet", "ModelModelEncoderPrenet", "ModelPreTrainedModel", "None", "__init__", "attention_mask", "class", "config", "def", "forward", "hidden_states", "input_values", "kwargs", "output_attentions", "output_hidden_states", "outputs", "post_init...
speecht5/modeling_speecht5.py:SpeechT5EncoderWithTextPrenet
[ 0, 0.04284796118736267, 0.022209152579307556, 0.022321321070194244, -0.00029093428747728467, 0.015030437149107456, 0.030733877792954445, -0.020863143727183342, -0.0009043499594554305, 0.007234799675643444, 0.029163533821702003, 0.03948293998837471, -0.0017946790903806686, 0.006786129903048...
[ "ModelEncoder", "ModelEncoderWithTextPrenet", "ModelPreTrainedModel", "ModelTextEncoderPrenet", "None", "__init__", "attention_mask", "class", "config", "def", "forward", "get_input_embeddings", "hidden_states", "input_values", "kwargs", "output_attentions", "output_hidden_states", ...
speecht5/modeling_speecht5.py:SpeechT5EncoderWithoutPrenet
[ -0.000023504424461862072, 0.04033707454800606, 0.020948536694049835, 0.016825685277581215, -0.00030120485462248325, 0.019277110695838928, 0.008747127838432789, -0.013092835433781147, 0.0021171392872929573, -0.0023260675370693207, -0.007354273460805416, 0.026408527046442032, 0.000363883329555...
[ "ModelEncoder", "ModelEncoderWithoutPrenet", "ModelPreTrainedModel", "None", "__init__", "attention_mask", "class", "config", "def", "forward", "hidden_states", "input_values", "kwargs", "output_attentions", "output_hidden_states", "post_init", "return", "return_dict", "self", ...
speecht5/modeling_speecht5.py:SpeechT5Decoder
[ -0.00009206173854181543, 0.0641942098736763, 0.016273008659482002, -0.00998825952410698, -0.0003945502976421267, 0.01739528402686119, 0.03883076459169388, -0.03703512251377106, 0.007407024037092924, -0.018966471776366234, 0.006705601699650288, 0.01941538229584694, -0.0023708089720457792, 0...
[ "BaseModelOutputWithPastAndCrossAttentions", "DynamicCache", "EncoderDecoderCache", "False", "ModelDecoder", "ModelDecoderLayer", "ModelPreTrainedModel", "ModuleList", "None", "Setting", "True", "__init__", "_prepare_4d_attention_mask", "_prepare_4d_causal_attention_mask", "all_cross_att...
speecht5/modeling_speecht5.py:SpeechT5DecoderWithSpeechPrenet
[ -0.00007860400364734232, 0.05890645459294319, -0.015963198617100716, 0.0007377358851954341, -0.0003811635251622647, 0.015063864178955555, 0.06070512533187866, -0.027204889804124832, 0.007644348777830601, -0.00474961381405592, 0.021696461364626884, 0.03035256266593933, -0.0025855887215584517,...
[ "ModelDecoder", "ModelDecoderWithModelPrenet", "ModelModelDecoderPrenet", "ModelPreTrainedModel", "None", "__init__", "attention_mask", "cache_position", "class", "config", "decoder_hidden_states", "def", "encoder_attention_mask", "encoder_hidden_states", "forward", "hidden_states", ...
speecht5/modeling_speecht5.py:SpeechT5DecoderWithTextPrenet
[ -0.0000662987440591678, 0.05552585795521736, 0.008935841731727123, 0.0010116046760231256, -0.000569027615711093, 0.0023182607255876064, 0.0445106066763401, -0.0370921716094017, 0.0024447112809866667, 0.002191810170188546, 0.022480104118585587, 0.0341697596013546, -0.003653016872704029, 0.0...
[ "ModelDecoder", "ModelDecoderWithTextPrenet", "ModelPreTrainedModel", "ModelTextDecoderPrenet", "None", "__init__", "attention_mask", "cache_position", "class", "config", "decoder_hidden_states", "def", "encoder_attention_mask", "encoder_hidden_states", "forward", "get_input_embeddings...
speecht5/modeling_speecht5.py:SpeechT5DecoderWithoutPrenet
[ -0.0000342068997269962, 0.054884422570466995, 0.007474098354578018, -0.006386449560523033, -0.00038869495620019734, 0.007641429081559181, 0.026661336421966553, -0.023091617971658707, 0.00758565217256546, -0.011936246417462826, -0.008199197240173817, 0.02164141833782196, -0.001268923399038612...
[ "ModelDecoder", "ModelDecoderWithoutPrenet", "ModelPreTrainedModel", "None", "__init__", "attention_mask", "cache_position", "class", "config", "def", "encoder_attention_mask", "encoder_hidden_states", "forward", "hidden_states", "input_values", "kwargs", "output_attentions", "outp...
speecht5/modeling_speecht5.py:SpeechT5GuidedMultiheadAttentionLoss
[ -0.00004744661782751791, 0.025867193937301636, 0.03351488709449768, -0.03059076890349388, -0.00030400982359424233, 0.043861765414476395, 0.018107036128640175, -0.026429524645209312, 0.012933596968650818, -0.0034583313390612602, 0.027554184198379517, 0.03239022567868233, 0.002474253298714757,...
[ "ModelGuidedMultiheadAttentionLoss", "Module", "__init__", "_make_guided_attention_mask", "_make_guided_attention_masks", "arange", "attentions", "class", "config", "def", "device", "enumerate", "exp", "float", "for", "forward", "grid_x", "grid_y", "guided_attention_loss_scale", ...
speecht5/modeling_speecht5.py:SpeechT5SpectrogramLoss
[ -0.00010445075167808682, 0.02651563100516796, 0.01611076295375824, -0.007132369093596935, -0.00019928679103031754, 0.03669673949480057, 0.016222642734646797, -0.012922174297273159, 0.01029298733919859, -0.020697856321930885, 0.04989861324429512, 0.040053147822618484, 0.0006642892840318382, ...
[ "BCEWithLogitsLoss", "L1Loss", "ModelGuidedMultiheadAttentionLoss", "ModelSpectrogramLoss", "Module", "None", "__init__", "attention_mask", "attn", "attn_criterion", "attn_loss", "bce_criterion", "bce_loss", "cat", "class", "config", "cross_attentions", "def", "device", "dim", ...
speecht5/modeling_speecht5.py:SpeechT5Model
[ 0.00008536216773791239, 0.055195003747940063, -0.005913750268518925, 0.009630965068936348, 0.0005491339834406972, 0.029512429609894753, 0.034919288009405136, -0.028723930940032005, 0.014418286271393299, -0.013911393471062183, 0.03221585974097252, 0.009124072268605232, -0.0010841875337064266,...
[ "BaseModelOutput", "ModelDecoderWithModelPrenet", "ModelDecoderWithTextPrenet", "ModelDecoderWithoutPrenet", "ModelEncoderWithModelPrenet", "ModelEncoderWithTextPrenet", "ModelEncoderWithoutPrenet", "ModelModel", "ModelPreTrainedModel", "None", "NotImplementedError", "Seq2SeqModelOutput", "_...
speecht5/modeling_speecht5.py:SpeechT5ForSpeechToText
[ -0.00004582770270644687, 0.06972096115350723, 0.006480474025011063, -0.004944154527038336, 0, 0.025028036907315254, 0.03150850906968117, -0.027821345254778862, 0.0009217915358021855, -0.004553091712296009, 0.035977803170681, 0.004581024404615164, -0.001173189259134233, 0.014748664572834969...
[ "CrossEntropyLoss", "GenerationMixin", "Model", "ModelDecoderWithTextPrenet", "ModelEncoderWithModelPrenet", "ModelForModeloText", "ModelModel", "ModelPreTrainedModel", "ModelTextDecoderPostnet", "Model_encoder", "None", "Please", "Seq2SeqLMOutput", "True", "ValueError", "You", "__cl...
speecht5/modeling_speecht5.py:_generate_speech
[ -0.00007710522913839668, 0.05059521272778511, -0.02416318468749523, 0.015201251953840256, -0.00047149404417723417, 0.02643202804028988, 0.08440098166465759, -0.00578555092215538, 0.015087809413671494, -0.01826419122517109, 0.052183400839567184, 0.01781042106449604, -0.003176380880177021, -...
[ "False", "Model", "ModelEncoderWithSpeechPrenet", "None", "True", "_generate_speech", "_get_feature_vector_attention_mask", "append", "attention_mask", "batch_first", "break", "bsz", "cat", "config", "continue", "cross_attentions", "decoder", "decoder_hidden_states", "decoder_out...
speecht5/modeling_speecht5.py:SpeechT5ForTextToSpeech
[ -0.00008045549475355074, 0.06482866406440735, -0.018570709973573685, -0.004811502061784267, -0.0001960827794391662, 0.033764928579330444, 0.050197191536426544, -0.021046804264187813, 0.0029966372530907393, -0.009679279290139675, 0.04051791504025459, 0.011142426170408726, -0.00416434090584516...
[ "False", "Model", "ModelDecoderWithModelPrenet", "ModelEncoderWithTextPrenet", "ModelForTextToModel", "ModelModel", "ModelModelDecoderPostnet", "ModelPreTrainedModel", "ModelSpectrogramLoss", "Model_decoder", "Model_decoder_postnet", "None", "Please", "Seq2SeqSpectrogramOutput", "The", ...
speecht5/modeling_speecht5.py:SpeechT5ForSpeechToSpeech
[ 0.000046596625907113776, 0.06256959587335587, -0.019243527203798294, 0.0043326071463525295, 0.00029892174643464386, 0.03556114062666893, 0.04253832623362541, -0.019806204363703728, 0.007708664517849684, -0.015529864467680454, 0.034660857170820236, 0.007877467200160027, -0.0002303455985384062...
[ "False", "Model", "ModelDecoderWithModelPrenet", "ModelEncoderWithModelPrenet", "ModelForModeloModel", "ModelModel", "ModelModelDecoderPostnet", "ModelPreTrainedModel", "Model_decoder", "Model_decoder_postnet", "Model_encoder", "None", "Seq2SeqSpectrogramOutput", "True", "_", "__init__...
speecht5/modeling_speecht5.py:HifiGanResidualBlock
[ -0.00004691328285844065, 0.012236400507390499, -0.018014701083302498, -0.024132901802659035, 0.0004000906483270228, 0.04645300284028053, 0.013029500842094421, -0.02152700163424015, 0.011613250710070133, -0.019034400582313538, 0.023226501420140266, -0.012123100459575653, 0.0012038125423714519...
[ "Conv1d", "ModelGanResidualBlock", "Module", "ModuleList", "_", "__init__", "apply_weight_norm", "channels", "class", "conv1", "conv2", "convs1", "convs2", "def", "dilation", "for", "forward", "functional", "get_padding", "hasattr", "hidden_states", "i", "if", "in", "...
speecht5/modeling_speecht5.py:SpeechT5HifiGan
[ -0.00028376519912853837, 0.03287366032600403, -0.021379372105002403, 0.013908086344599724, -0.0011135089443996549, -0.010517272166907787, 0.058161091059446335, -0.02666674368083477, 0.0018247179687023163, -0.0317242294549942, 0.0349426306784153, 0.002025868045166135, -0.0035632287617772818, ...
[ "Conv1d", "ConvTranspose1d", "HifiGanResidualBlock", "ModelHifiGan", "ModelHifiGanConfig", "ModuleList", "PreTrainedModel", "__init__", "_init_weights", "append", "apply_weight_norm", "auto_docstring", "channels", "class", "config", "conv_post", "conv_pre", "custom_intro", "def",...
gpt_oss/modeling_gpt_oss.py:GptOssRMSNorm
[ -0.00008900701504899189, 0.041736360639333725, 0.02910265140235424, 0.051888447254896164, -0.0003807032771874219, 0.039029136300086975, 0.01985297165811062, -0.026169825345277786, 0.007162861526012421, 0.04399237781763077, 0.01827375777065754, 0.003750632284209132, 0.002749523613601923, 0....
[ "ModelRMSNorm", "Module", "Parameter", "True", "__init__", "class", "def", "eps", "extra_repr", "f", "float32", "forward", "hidden_size", "hidden_states", "keepdim", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "shape", "super", "to", "torch", "tuple", ...
gpt_oss/modeling_gpt_oss.py:GptOssExperts
[ -0.00040273863123729825, 0.02094976417720318, -0.02412753738462925, 0.004590116906911135, -0.0016771580558270216, 0.06402624398469925, 0.05555218458175659, -0.008003280498087406, -0.006855751387774944, -0.014711912721395493, 0.02118515409529209, -0.026481444016098976, -0.0010004100622609258,...
[ "ModelExperts", "Module", "None", "Parameter", "__init__", "_apply_gate", "alpha", "clamp", "class", "config", "continue", "current_state", "def", "device", "dim", "down_proj", "down_proj_bias", "dtype", "empty", "expert_hit", "expert_idx", "expert_mask", "for", "forwar...
gpt_oss/modeling_gpt_oss.py:GptOssTopKRouter
[ -0.0003609287668950856, 0.03470779210329056, 0.003590966574847698, -0.010025392286479473, -0.0016342562157660723, 0.05651741474866867, 0.07645094394683838, -0.020637065172195435, -0.0018101402092725039, 0.012077372521162033, 0.022044137120246887, -0.02028529718518257, -0.0009820193517953157,...
[ "F", "ModelTopKRouter", "Module", "Parameter", "__init__", "bias", "class", "config", "def", "dim", "dtype", "forward", "functional", "hidden_dim", "hidden_size", "hidden_states", "linear", "nn", "num_experts", "num_experts_per_tok", "num_local_experts", "return", "router...
gpt_oss/modeling_gpt_oss.py:GptOssMLP
[ -0.00046644036774523556, 0.025672877207398415, 0.006627184804528952, 0.021135346964001656, -0.0016642592381685972, 0.07880976796150208, 0.048479948192834854, -0.026986373588442802, -0.002686696592718363, -0.0017388897249475121, 0.023284703493118286, -0.03725552558898926, -0.00021922697487752...
[ "ModelExperts", "ModelMLP", "ModelTopKRouter", "Module", "_", "__init__", "batch_size", "class", "config", "def", "experts", "forward", "hidden_dim", "hidden_states", "nn", "reshape", "return", "router", "router_indices", "router_scores", "self", "sequence_length", "shape...
gpt_oss/modeling_gpt_oss.py:GptOssRotaryEmbedding
[ -0.0003310962056275457, 0.047736067324876785, 0.0015063057653605938, -0.005326646380126476, -0.0016154584009200335, 0.043311748653650284, 0.04168173670768738, 0.0038421712815761566, -0.0018919783178716898, 0.022121591493487358, 0.0006440003053285182, -0.0046571772545576096, -0.00129527738317...
[ "False", "ModelRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "arange", "attention_factor", "attention_scaling", "base", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cpu", "def", "default", "device", "device...
gpt_oss/modeling_gpt_oss.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
gpt_oss/modeling_gpt_oss.py:_apply_rotary_emb
[ -0.00011168773198733106, 0.023619528859853745, 0.03322555497288704, -0.0025851516984403133, -0.0006957307341508567, 0.03503375127911568, 0.04136242717504501, -0.015256633050739765, 0.01706482656300068, 0.001652801875025034, 0.010679642669856548, 0.013165908865630627, -0.00015892325609456748,...
[ "_apply_rotary_emb", "cat", "chunk", "cos", "def", "dim", "first_", "first_half", "return", "second_", "second_half", "sin", "torch", "x" ]
gpt_oss/modeling_gpt_oss.py:apply_rotary_pos_emb
[ -0.00021975146955810487, 0.026865288615226746, 0.01607324928045273, -0.0035447257105261087, -0.0009830514900386333, 0.032835353165864944, 0.058782171458005905, -0.007519984617829323, 0.01653248630464077, 0.025143155828118324, 0.001894347253255546, 0.0030998410657048225, -0.000315724551910534...
[ "Model_rotary_pos_emb", "None", "_Model_rotary_emb", "cos", "def", "k", "k_embed", "position_ids", "q", "q_embed", "return", "sin", "unsqueeze", "unsqueeze_dim" ]
gpt_oss/modeling_gpt_oss.py:eager_attention_forward
[ -0.00008625836198916659, 0.02356187254190445, -0.0054067098535597324, -0.013203755021095276, -0.0002774495806079358, 0.040066566318273544, 0.0610104538500309, -0.03687945380806923, 0.023675698786973953, 0.011496372520923615, 0.032326433807611465, 0.024813951924443245, 0.0015579861355945468, ...
[ "F", "Model_attention_forward", "None", "True", "attention_mask", "attn_output", "attn_weights", "cat", "causal_mask", "combined_logits", "contiguous", "def", "dim", "dropout", "dtype", "expand", "functional", "if", "is", "keepdim", "key", "key_states", "kwargs", "matmu...
gpt_oss/modeling_gpt_oss.py:GptOssAttention
[ -0.000058491779782343656, 0.03219950571656227, 0.017563367262482643, -0.004278256092220545, -0.0004995989729650319, 0.030172962695360184, 0.04390841722488403, -0.013059939257800579, 0.002603543922305107, 0.004165670368820429, 0.02049059420824051, 0.027020564302802086, -0.0008795756730251014,...
[ "ALL_ATTENTION_FUNCTIONS", "Linear", "ModelAttention", "Module", "None", "Parameter", "Tensor", "True", "__init__", "_attn_implementation", "apply_rotary_pos_emb", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "cache_kwargs", "cache_po...
gpt_oss/modeling_gpt_oss.py:GptOssDecoderLayer
[ -0.00020187246263958514, 0.04329278692603111, 0.013146501034498215, -0.0009349882020615041, -0.000768531986977905, 0.04125281423330307, 0.04329278692603111, -0.0330929160118103, 0.006799914408475161, 0.0007968649733811617, -0.004391611088067293, 0.018586432561278343, -0.0011616520350798965, ...
[ "False", "GradientCheckpointingLayer", "ModelAttention", "ModelDecoderLayer", "ModelMLP", "ModelRMSNorm", "None", "Tensor", "_", "__init__", "attention_mask", "attention_type", "cache_position", "class", "config", "def", "eps", "forward", "hidden_size", "hidden_states", "inpu...