identifier
stringlengths
24
102
embedding
listlengths
2.56k
2.56k
tokens
listlengths
4
448
focalnet/modeling_focalnet.py:FocalNetModel
[ -0.000033531279768794775, 0.03186742961406708, -0.001423654961399734, 0.03052091784775257, 0, 0.03859998285770416, 0.015148249454796314, -0.01604592427611351, 0.008303484879434109, 0.017953481525182724, 0.029398825019598007, -0.0015428772894665599, -0.00045585009502246976, -0.0036748531274...
[ "AdaptiveAvgPool1d", "False", "LayerNorm", "ModelEmbeddings", "ModelEncoder", "ModelModel", "ModelModelOutput", "ModelPreTrainedModel", "None", "True", "__init__", "add_pooling_layer", "auto_docstring", "bool_masked_pos", "class", "config", "def", "depths", "else", "embed_dim",...
focalnet/modeling_focalnet.py:FocalNetForMaskedImageModeling
[ -0.0001356009888695553, 0.05487028509378433, 0.011701928451657295, 0.00306545733474195, -0.0005773977609351277, 0.03381801024079323, 0.005319057963788509, -0.06181305646896362, 0.00974227488040924, 0.01534128375351429, 0.013885541819036007, 0.014669402502477169, -0.001217784360051155, 0.00...
[ "Conv2d", "False", "Model", "ModelForMaskedImageModeling", "ModelMaskedImageModelingOutput", "ModelModel", "ModelPreTrainedModel", "None", "PixelShuffle", "Sequential", "True", "__init__", "add_pooling_layer", "auto_docstring", "batch_size", "bool_masked_pos", "class", "config", ...
focalnet/modeling_focalnet.py:FocalNetForImageClassification
[ -0.00008442571561317891, 0.036095909774303436, 0.011419231072068214, 0.029188668355345726, -0.00009008311462821439, 0.0011488860473036766, 0.020164690911769867, -0.007575685158371925, 0.0037878425791859627, 0.011697748675942421, 0.04589973762631416, 0.0012533302651718259, 0.00044562853872776...
[ "Identity", "Linear", "Model", "ModelForImageClassification", "ModelImageClassifierOutput", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "auto_docstring", "class", "classifier", "config", "def", "else", "forward", "hidden_states", "if", "is", "kwargs", "labels", ...
focalnet/modeling_focalnet.py:FocalNetBackbone
[ -0.00021211968851275742, -0.004543919116258621, 0.030292794108390808, 0.02412204071879387, -0.0006696670898236334, 0.02142934687435627, 0.003702452639117837, -0.015370788052678108, -0.0013463464565575123, 0.0011289676185697317, -0.00992930494248867, -0.0008835398475639522, -0.001135979779064...
[ "BackboneMixin", "BackboneOutput", "False", "Model", "ModelBackbone", "ModelModel", "ModelPreTrainedModel", "None", "True", "__init__", "_init_backbone", "attentions", "auto_docstring", "class", "config", "def", "else", "embed_dim", "enumerate", "feature_maps", "for", "forw...
albert/modeling_albert.py:AlbertEmbeddings
[ -0.00024960190057754517, 0.018713010475039482, 0.00941355712711811, -0.008215468376874924, -0.0012052205856889486, 0.0381106436252594, 0.031036213040351868, -0.004421519115567207, 0.007045904640108347, -0.012323201633989811, 0.02076687663793564, 0.022820744663476944, -0.0015118742594495416, ...
[ "Dropout", "Embedding", "False", "LayerNorm", "ModelEmbeddings", "Module", "None", "__init__", "arange", "batch_size", "buffered_token_type_ids", "class", "config", "def", "device", "dim", "dropout", "dtype", "else", "embedding_size", "embeddings", "eps", "expand", "for...
albert/modeling_albert.py:eager_attention_forward
[ 0.00002095530362566933, 0.025862814858555794, 0.0269921962171793, -0.01146321278065443, 0.00008955634984886274, 0.03704368323087692, 0.060083046555519104, -0.023942869156599045, 0.021458230912685394, 0.014569009654223919, 0.023152301087975502, 0.026879258453845978, 0.003063444746658206, -0...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "shape", "size", "softmax", "tor...
albert/modeling_albert.py:AlbertAttention
[ -0.00010192796617047861, 0.03261695057153702, 0.04813811928033829, -0.00009358041279483587, -0.0003866233164444566, 0.030367504805326462, 0.03869045153260231, -0.014059029519557953, 0.0027274517342448235, 0.02316928096115589, 0.018108030781149864, 0.010797334834933281, -0.0016308474587276578...
[ "ALL_ATTENTION_FUNCTIONS", "Dropout", "False", "LayerNorm", "Linear", "ModelAttention", "Module", "None", "The", "ValueError", "__init__", "_attn_implementation", "a", "all_head_size", "and", "attention", "attention_dropout", "attention_head_size", "attention_interface", "atten...
albert/modeling_albert.py:AlbertLayer
[ -0.00013687126920558512, 0.027013972401618958, 0.024527329951524734, 0.018988901749253273, -0.000196034976397641, 0.032100286334753036, 0.03390875086188316, -0.018762841820716858, 0.004153822083026171, 0.009663994424045086, 0.01921495981514454, -0.0018225953681394458, 0.00372996274381876, ...
[ "ACT2FN", "Dropout", "LayerNorm", "Linear", "ModelAttention", "ModelLayer", "Module", "None", "_", "__init__", "activation", "apply_chunking_to_forward", "attention", "attention_mask", "attention_output", "chunk_size_feed_forward", "class", "config", "def", "dropout", "eps", ...
albert/modeling_albert.py:AlbertLayerGroup
[ 0.00006477566785179079, -0.018967734649777412, -0.009427078068256378, 0.022602271288633347, 0.000553698861040175, 0.060424163937568665, 0.0177183635532856, -0.017604785040020943, 0.011982611380517483, -0.026123227551579475, 0.019422052428126335, -0.02896270900964737, -0.0009299301309511065, ...
[ "ModelLayer", "ModelLayerGroup", "Model_layer", "Model_layers", "Module", "ModuleList", "None", "_", "__init__", "attention_mask", "class", "config", "def", "enumerate", "for", "forward", "hidden_states", "in", "inner_group_num", "kwargs", "layer_index", "nn", "range", ...
albert/modeling_albert.py:AlbertTransformer
[ -0.00008017026993911713, 0.012897336855530739, 0.021308643743395805, 0.007149610668420792, -0.00019538764900062233, 0.039252765476703644, 0.010149643756449223, -0.02321520633995533, 0.0018364686984568834, -0.03409383073449135, 0.011831904761493206, -0.014299221336841583, -0.00142992218025028...
[ "BaseModelOutput", "Linear", "ModelLayerGroup", "ModelTransformer", "Model_layer_groups", "Module", "ModuleList", "None", "_", "__init__", "attention_mask", "class", "config", "def", "embedding_hidden_mapping_in", "embedding_size", "for", "forward", "group_idx", "hidden_size", ...
albert/modeling_albert.py:AlbertPreTrainedModel
[ -0.00016960255743470043, 0.03256369009613991, 0.0013214866630733013, 0.015151161700487137, -0.0009681479423306882, 0.031206870451569557, 0.023744357749819756, -0.016960255801677704, 0.002501637674868107, -0.004777138587087393, 0.006190493702888489, 0.014359683729708195, -0.00108828314114362,...
[ "Embedding", "False", "LayerNorm", "Linear", "Model", "ModelAttention", "ModelConfig", "ModelEmbeddings", "ModelLayer", "ModelMLMHead", "ModelPreTrainedModel", "None", "PreTrainedModel", "True", "_can_record_outputs", "_init_weights", "_is_hf_initialized", "_supports_attention_back...
albert/modeling_albert.py:AlbertForPreTrainingOutput
[ -0.00027908937772735953, 0.05037920922040939, 0.01625874452292919, 0.023472130298614502, -0.001152138109318912, 0.03961637616157532, 0.06411898881196976, -0.01694573275744915, 0.011678816750645638, -0.01236580591648817, 0.018434209749102592, 0.02198365516960621, -0.0018033466767519712, -0....
[ "ModelForPreTrainingOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "loss", "prediction_logits", "r", "sop_logits" ]
albert/modeling_albert.py:AlbertModel
[ 0.000053764175390824676, 0.019134800881147385, 0.013204132206737995, 0.011861338280141354, 0.00027450325433164835, 0.03536021709442139, 0.014546925202012062, 0.0077210599556565285, 0.004727750550955534, -0.003342995187267661, 0.015442119911313057, -0.012756533920764923, -0.000243031550780870...
[ "BaseModelOutputWithPooling", "Linear", "Model", "ModelConfig", "ModelEmbeddings", "ModelModel", "ModelPreTrainedModel", "ModelTransformer", "None", "Tanh", "True", "ValueError", "You", "__init__", "_attn_implementation", "add_pooling_layer", "attention_mask", "attn_implementation"...
albert/modeling_albert.py:AlbertForPreTraining
[ -0.00017641477461438626, 0.046510305255651474, 0.014941716566681862, 0.019772497937083244, -0.0007512987358495593, 0.039095621556043625, 0.020334215834736824, -0.007246171124279499, 0.0024434763472527266, 0.006066561676561832, 0.024603277444839478, 0.02314280904829502, -0.0008706639637239277...
[ "CrossEntropyLoss", "Model", "ModelForPreTraining", "ModelForPreTrainingOutput", "ModelMLMHead", "ModelModel", "ModelPreTrainedModel", "ModelSOPHead", "None", "True", "__init__", "_tied_weights_keys", "and", "attention_mask", "attentions", "auto_docstring", "bias", "can_return_tupl...
albert/modeling_albert.py:AlbertMLMHead
[ -0.0003810608177445829, 0.03925289213657379, 0.031588125973939896, 0.03693023696541786, -0.0015895679825916886, 0.033910784870386124, 0.05040164664387703, -0.014632735401391983, 0.0034114017616957426, 0.007722832728177309, 0.01683925837278366, 0.010510020889341831, -0.00035928591387346387, ...
[ "ACT2FN", "LayerNorm", "Linear", "ModelMLMHead", "Module", "Parameter", "__init__", "activation", "bias", "class", "config", "decoder", "def", "dense", "embedding_size", "eps", "forward", "hidden_act", "hidden_size", "hidden_states", "layer_norm_eps", "nn", "prediction_sc...
albert/modeling_albert.py:AlbertSOPHead
[ -0.0003381282731425017, 0.042354632169008255, 0.022913163527846336, 0.01539116445928812, -0.0014465380227193236, 0.025806238874793053, 0.05369549244642258, 0.011051550507545471, -0.005005021579563618, 0.01359745766967535, 0.0319395586848259, 0.0017864744877442718, -0.00041407151729799807, ...
[ "Dropout", "Linear", "ModelSOPHead", "Module", "__init__", "class", "classifier", "classifier_dropout_prob", "config", "def", "dropout", "dropout_pooled_output", "forward", "hidden_size", "logits", "nn", "num_labels", "pooled_output", "return", "self", "super" ]
albert/modeling_albert.py:AlbertForMaskedLM
[ -0.00008456408249912784, 0.036824606359004974, 0.017742764204740524, -0.01450666319578886, -0.0005858460208401084, 0.04218091070652008, 0.01751958578824997, -0.01316758617758751, 0.005942152347415686, -0.0005614357651211321, 0.021759994328022003, 0.0357087068259716, -0.0002964101731777191, ...
[ "CrossEntropyLoss", "False", "MaskedLMOutput", "Model", "ModelForMaskedLM", "ModelMLMHead", "ModelModel", "ModelPreTrainedModel", "None", "True", "__init__", "_tied_weights_keys", "add_pooling_layer", "attention_mask", "attentions", "auto_docstring", "bias", "can_return_tuple", "...
albert/modeling_albert.py:AlbertForSequenceClassification
[ -0.00038339142338372767, 0.02995132841169834, 0.009618984535336494, 0.010655761696398258, -0.0013607695000246167, 0.03133369982242584, 0.023961063474416733, 0.011922933161258698, -0.003427123185247183, 0.013478098437190056, 0.0550643652677536, -0.0010511764558032155, 0, -0.0179707966744899...
[ "BCEWithLogitsLoss", "CrossEntropyLoss", "Dropout", "Linear", "MSELoss", "Model", "ModelForSequenceClassification", "ModelModel", "ModelPreTrainedModel", "None", "SequenceClassifierOutput", "True", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "can_return_tup...
albert/modeling_albert.py:AlbertForTokenClassification
[ -0.00026531689218245447, 0.03464433178305626, 0.008319198153913021, -0.02803455851972103, -0.0010398997692391276, 0.03943072259426117, 0.041254106909036636, 0.015954626724123955, -0.0019231024198234081, 0.010997296310961246, 0.05173857510089874, 0.020627053454518318, -0.0006267889402806759, ...
[ "CrossEntropyLoss", "Dropout", "False", "Linear", "Model", "ModelForTokenClassification", "ModelModel", "ModelPreTrainedModel", "None", "TokenClassifierOutput", "True", "__init__", "add_pooling_layer", "attention_mask", "attentions", "auto_docstring", "can_return_tuple", "class", ...
albert/modeling_albert.py:AlbertForQuestionAnswering
[ -0.0001946567208506167, 0.026719937101006508, 0.01341633964329958, 0.01578393019735813, -0.000739871698897332, 0.05118502676486969, 0.03991079330444336, 0.02345040999352932, 0.005890788044780493, 0.032695282250642776, 0.0216465312987566, 0.018940715119242668, 0.0010358203435316682, -0.0012...
[ "CrossEntropyLoss", "False", "Linear", "Model", "ModelForQuestionAnswering", "ModelModel", "ModelPreTrainedModel", "None", "QuestionAnsweringModelOutput", "True", "__init__", "add_pooling_layer", "and", "attention_mask", "attentions", "auto_docstring", "can_return_tuple", "clamp", ...
albert/modeling_albert.py:AlbertForMultipleChoice
[ -0.00029660252039320767, 0.05246089771389961, 0.01898256130516529, 0.026805676519870758, -0.0013589787995442748, 0.043947506695985794, 0.049239613115787506, -0.005205823574215174, 0.002056444063782692, 0.01409311406314373, 0.039575763046741486, -0.008110730908811092, -0.0013158366782590747, ...
[ "CrossEntropyLoss", "Dropout", "Linear", "Model", "ModelForMultipleChoice", "ModelModel", "ModelPreTrainedModel", "MultipleChoiceModelOutput", "None", "True", "__init__", "attention_mask", "attentions", "auto_docstring", "can_return_tuple", "class", "classifier", "classifier_dropou...
electra/modeling_electra.py:ElectraEmbeddings
[ -0.000249445263762027, 0.017332881689071655, 0.008096280507743359, -0.009521681815385818, -0.0012115912977606058, 0.034209635108709335, 0.03010447882115841, -0.01402595080435276, 0.006585354916751385, -0.012144421227276325, 0.023034488782286644, 0.027367709204554558, -0.00104767014272511, ...
[ "Dropout", "Embedding", "False", "LayerNorm", "ModelEmbeddings", "Module", "None", "__init__", "arange", "batch_size", "buffered_token_type_ids", "class", "config", "def", "device", "dim", "dropout", "dtype", "else", "embedding_size", "embeddings", "eps", "expand", "for...
electra/modeling_electra.py:eager_attention_forward
[ 0.00002095530362566933, 0.025862814858555794, 0.0269921962171793, -0.01146321278065443, 0.00008955634984886274, 0.03704368323087692, 0.060083046555519104, -0.023942869156599045, 0.021458230912685394, 0.014569009654223919, 0.023152301087975502, 0.026879258453845978, 0.003063444746658206, -0...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "shape", "size", "softmax", "tor...
electra/modeling_electra.py:ElectraSelfAttention
[ -0.00004590151002048515, 0.03693104162812233, 0.04006458446383476, -0.006714735180139542, -0.00029376966995187104, 0.008001726120710373, 0.023389659821987152, -0.027866151183843613, 0.0027698283083736897, 0.030216308310627937, 0.010239970870316029, 0.007945769466459751, -0.000088742919615469...
[ "ALL_ATTENTION_FUNCTIONS", "Dropout", "EncoderDecoderCache", "False", "Linear", "ModelSelfAttention", "Module", "None", "The", "ValueError", "__init__", "_attn_implementation", "a", "all_head_size", "and", "attention", "attention_head_size", "attention_interface", "attention_mask...
electra/modeling_electra.py:ElectraCrossAttention
[ -0.00026539957616478205, 0.03879137337207794, 0.05462927371263504, -0.0018291050801053643, -0.0012409222545102239, 0.020658127963542938, 0.029839517548680305, -0.03190533071756363, -0.0006957772420719266, 0.022953474894165993, 0.007689414545893669, 0.013026097789406776, -0.000245673931203782...
[ "ALL_ATTENTION_FUNCTIONS", "Dropout", "False", "Linear", "ModelCrossAttention", "Module", "None", "The", "True", "ValueError", "__init__", "_attn_implementation", "a", "all_head_size", "and", "attention", "attention_head_size", "attention_interface", "attention_mask", "attentio...
electra/modeling_electra.py:ElectraSelfOutput
[ -0.00012168083776487038, 0.04875698313117027, 0.03859927877783775, 0.020879726856946945, -0.0006630723946727812, 0.05620596557855606, 0.02370131015777588, -0.01952536590397358, 0.0035551965702325106, 0.017155233770608902, 0.01365646906197071, 0.003512872848659754, 0.0036962758749723434, -0...
[ "Dropout", "LayerNorm", "Linear", "ModelSelfOutput", "Module", "__init__", "class", "config", "def", "dense", "dropout", "eps", "forward", "hidden_dropout_prob", "hidden_size", "hidden_states", "input_tensor", "layer_norm_eps", "nn", "return", "self", "super" ]
electra/modeling_electra.py:ElectraAttention
[ 0.00006504472548840567, 0.037578269839286804, 0.03577811270952225, -0.0018423479050397873, 0.00018634434673003852, 0.0378032885491848, 0.03847834840416908, -0.017101489007472992, 0.011982292868196964, 0.008494488894939423, 0.0189016442745924, 0.03442799672484398, 0.0014696591533720493, -0....
[ "False", "ModelAttention", "ModelCrossAttention", "ModelSelfAttention", "ModelSelfOutput", "Module", "None", "__init__", "attention_class", "attention_mask", "attention_output", "attn_weights", "cache_position", "class", "config", "def", "else", "encoder_attention_mask", "encoder...
electra/modeling_electra.py:ElectraIntermediate
[ -0.00025367451598867774, 0.02240910567343235, 0.04047359153628349, 0.012862369418144226, -0.0009432404185645282, 0.03635763004422188, 0.03452831506729126, -0.018864808604121208, -0.001329111517407, -0.003772961674258113, 0.023209432139992714, -0.02103712037205696, -0.0013719861162826419, 0...
[ "ACT2FN", "Linear", "ModelIntermediate", "Module", "__init__", "class", "config", "def", "dense", "else", "forward", "hidden_act", "hidden_size", "hidden_states", "if", "intermediate_act_fn", "intermediate_size", "isinstance", "nn", "return", "self", "str", "super" ]
electra/modeling_electra.py:ElectraOutput
[ -0.0002591986849438399, 0.03935529664158821, 0.0507957898080349, 0.030660521239042282, -0.0012513039400801063, 0.04988054931163788, 0.03798243775963783, -0.023681821301579475, 0.002259497530758381, 0.014758236706256866, 0.011040075682103634, 0.00267421524040401, 0.0013013561256229877, 0.00...
[ "Dropout", "LayerNorm", "Linear", "ModelOutput", "Module", "__init__", "class", "config", "def", "dense", "dropout", "eps", "forward", "hidden_dropout_prob", "hidden_size", "hidden_states", "input_tensor", "intermediate_size", "layer_norm_eps", "nn", "return", "self", "su...
electra/modeling_electra.py:ElectraLayer
[ -0.00008142412843881175, 0.023534199222922325, 0.0189394261687994, 0.003950383514165878, -0.00009981021867133677, 0.03832712396979332, 0.02062044106423855, -0.013728282414376736, 0.003502112813293934, 0.005435279104858637, 0.0047628735192120075, 0.010814524255692959, 0.002661605831235647, ...
[ "False", "GradientCheckpointingLayer", "If", "ModelAttention", "ModelIntermediate", "ModelLayer", "ModelOutput", "None", "True", "ValueError", "_", "__init__", "add_cross_attention", "and", "apply_chunking_to_forward", "are", "attention", "attention_mask", "attention_output", "...
electra/modeling_electra.py:ElectraEncoder
[ -0.000011508945135574322, 0.014999293722212315, 0.0157887302339077, 0.01770593412220478, -0.00019647806766442955, 0.04894506558775902, 0.020976455882191658, -0.03699073940515518, 0.011785159818828106, 0.012067100964486599, 0.010995723307132721, 0.0013674168149009347, -0.0004546308482531458, ...
[ "BaseModelOutputWithPastAndCrossAttentions", "ModelEncoder", "ModelLayer", "Module", "ModuleList", "None", "__init__", "attention_mask", "cache_position", "class", "config", "def", "else", "encoder_attention_mask", "encoder_hidden_states", "enumerate", "for", "forward", "hidden_s...
electra/modeling_electra.py:ElectraDiscriminatorPredictions
[ -0.00018399152031634003, 0.049366340041160583, 0.020607050508260727, 0.0019956002943217754, -0.0009765704162418842, 0.027853485196828842, 0.04008184373378754, -0.012737874872982502, 0.002080519450828433, 0.02638155408203602, 0.04529022052884102, 0.017210284247994423, 0.0019106811378151178, ...
[ "Linear", "ModelDiscriminatorPredictions", "Module", "__init__", "activation", "class", "config", "def", "dense", "dense_prediction", "discriminator_hidden_states", "forward", "get_activation", "hidden_act", "hidden_size", "hidden_states", "logits", "nn", "return", "self", "s...
electra/modeling_electra.py:ElectraGeneratorPredictions
[ -0.000295931997243315, 0.051044661551713943, 0.014262478798627853, 0.055202145129442215, -0.0013064314844086766, 0.020094504579901695, 0.048273004591464996, -0.00640945415943861, -0.0006568246753886342, 0, 0.021133875474333763, 0.010797908529639244, -0.0014507885789498687, 0.00186220626346...
[ "LayerNorm", "Linear", "ModelGeneratorPredictions", "Module", "__init__", "activation", "class", "config", "def", "dense", "embedding_size", "eps", "forward", "gelu", "generator_hidden_states", "get_activation", "hidden_size", "hidden_states", "layer_norm_eps", "nn", "return"...
electra/modeling_electra.py:ElectraPreTrainedModel
[ -0.00023997374228201807, 0.027986271306872368, -0.0029294572304934263, 0.0003786252345889807, -0.0010736603289842606, 0.03208182379603386, 0.030716639012098312, -0.010466409847140312, -0.002076217206194997, 0.0013011909322813153, 0.0031285465229302645, 0.012002241797745228, -0.00177758326753...
[ "Model", "ModelConfig", "ModelCrossAttention", "ModelEmbeddings", "ModelLayer", "ModelPreTrainedModel", "ModelSelfAttention", "PreTrainedModel", "True", "_can_record_outputs", "_init_weights", "_supports_attention_backend", "_supports_flash_attn", "_supports_flex_attn", "_supports_sdpa",...
electra/modeling_electra.py:ElectraForPreTrainingOutput
[ -0.00023770524421706796, 0.029518023133277893, 0.028609776869416237, 0.018051406368613243, -0.0010643518762663007, 0.05086182430386543, 0.051543012261390686, -0.004739913623780012, 0.0133398761972785, -0.027020344510674477, 0.006953765172511339, 0.020662616938352585, -0.0018448764458298683, ...
[ "ModelForPreTrainingOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "logits", "loss", "r" ]
electra/modeling_electra.py:ElectraModel
[ 0.000033400760003132746, 0.03621428459882736, 0.026713622733950615, -0.00314360111951828, 0, 0.038226187229156494, -0.000768435827922076, -0.025595897808670998, 0.007879959419369698, 0, 0.01318915281444788, 0.007935846224427223, 0.0025148808490484953, 0.01799536868929863, -0.022019177675...
[ "BaseModelOutputWithPastAndCrossAttentions", "DynamicCache", "EncoderDecoderCache", "False", "Linear", "ModelEmbeddings", "ModelEncoder", "ModelModel", "ModelPreTrainedModel", "None", "ValueError", "You", "__init__", "_create_attention_masks", "and", "arange", "attention_mask", "au...
electra/modeling_electra.py:ElectraClassificationHead
[ -0.0003039397415705025, 0.0259743332862854, 0.027004152536392212, 0.026088757440447807, -0.0009797586826607585, 0.013216015882790089, 0.052177514880895615, 0, -0.003947641234844923, 0.0057212188839912415, 0.025516636669635773, 0.003890428924933076, -0.0002145457110600546, -0.00652218936011...
[ "Dropout", "Linear", "ModelClassificationHead", "Module", "None", "__init__", "activation", "class", "classifier_dropout", "config", "def", "dense", "dropout", "else", "features", "forward", "gelu", "get_activation", "hidden_dropout_prob", "hidden_size", "if", "is", "kwar...
electra/modeling_electra.py:ElectraSequenceSummary
[ -0.00037931749830022454, 0.022525623440742493, 0.020658213645219803, 0.04575152322649956, -0.0016193939372897148, 0.025326736271381378, 0.020074648782610893, -0.01207980327308178, -0.0034576246980577707, 0.00682771485298872, 0.02917826734483242, 0.020658213645219803, 0.0016704559093341231, ...
[ "Dropout", "Identity", "Linear", "ModelSequenceSummary", "Module", "None", "NotImplementedError", "__init__", "activation", "activation_string", "and", "attn", "class", "cls_index", "config", "def", "dim", "dtype", "elif", "else", "expand", "first", "first_dropout", "fo...
electra/modeling_electra.py:ElectraForSequenceClassification
[ -0.0003659949288703501, 0.03306872025132179, -0.0031863090116530657, 0.008267180062830448, -0.0013276287354528904, 0.033528007566928864, 0.029279595240950584, 0.011941482312977314, -0.002368202665820718, 0.022505100816488266, 0.055344175547361374, 0.001062102965079248, -0.00042161182500422, ...
[ "BCEWithLogitsLoss", "CrossEntropyLoss", "MSELoss", "Model", "ModelClassificationHead", "ModelForSequenceClassification", "ModelModel", "ModelPreTrainedModel", "None", "SequenceClassifierOutput", "True", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "can_return...
electra/modeling_electra.py:ElectraForPreTraining
[ -0.00023282665642909706, 0.056894369423389435, 0.015013792552053928, -0.0011006351560354233, -0.0009454173850826919, 0.04402540624141693, 0.023141559213399887, 0, 0.004458983428776264, 0.016255535185337067, 0.04357386380434036, 0.00903085246682167, 0.0015310116577893496, -0.028221413493156...
[ "BCEWithLogitsLoss", "Model", "ModelDiscriminatorPredictions", "ModelForPreTraining", "ModelForPreTrainingOutput", "ModelModel", "ModelPreTrainedModel", "None", "True", "__init__", "active_labels", "active_logits", "active_loss", "attention_mask", "attentions", "auto_docstring", "can...
electra/modeling_electra.py:ElectraForMaskedLM
[ -0.00014977513637859374, 0.04686297848820686, 0.01278081163764, -0.013285317458212376, -0.0007357375579886138, 0.03923933580517769, 0.023655712604522705, 0.004204214550554752, 0.00560561940073967, -0.0035035121254622936, 0.014798834919929504, 0.031615693122148514, -0.0009809833718463778, -...
[ "CrossEntropyLoss", "Linear", "MaskedLMOutput", "Model", "ModelForMaskedLM", "ModelGeneratorPredictions", "ModelModel", "ModelPreTrainedModel", "None", "True", "__init__", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "can_return_tuple", "class", "config",...
electra/modeling_electra.py:ElectraForTokenClassification
[ -0.0002562478184700012, 0.0394052192568779, 0.00882631354033947, -0.03348304703831673, -0.0009538113372400403, 0.04282185807824135, 0.04396073892712593, 0.014008213765919209, -0.000861277396325022, 0.022777583450078964, 0.051932889968156815, 0.019930386915802956, -0.0004519926733337343, -0...
[ "CrossEntropyLoss", "Dropout", "Linear", "Model", "ModelForTokenClassification", "ModelModel", "ModelPreTrainedModel", "None", "TokenClassifierOutput", "True", "__init__", "attention_mask", "attentions", "auto_docstring", "can_return_tuple", "class", "classifier", "classifier_dropo...
electra/modeling_electra.py:ElectraForQuestionAnswering
[ -0.00021580536849796772, 0.029990779235959053, 0.015108137391507626, 0.00986538827419281, -0.0007821843610145152, 0.05051078647375107, 0.044873423874378204, 0.01973077654838562, 0.0028045889921486378, 0.03968704864382744, 0.026270119473338127, 0.014657148160040379, 0.00012155567674199119, ...
[ "CrossEntropyLoss", "Linear", "Model", "ModelConfig", "ModelForQuestionAnswering", "ModelModel", "ModelPreTrainedModel", "None", "QuestionAnsweringModelOutput", "True", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "base_model_prefix", "can_return_tuple", "cl...
electra/modeling_electra.py:ElectraForMultipleChoice
[ -0.0002744243247434497, 0.05268947035074234, 0.018475528806447983, 0.030564453452825546, -0.0009765229187905788, 0.04310956597328186, 0.04561858996748924, -0.008211346343159676, 0.003449905663728714, 0.019045760855078697, 0.03740724176168442, -0.008838601410388947, -0.0016109064454212785, ...
[ "CrossEntropyLoss", "Linear", "Model", "ModelForMultipleChoice", "ModelModel", "ModelPreTrainedModel", "ModelSequenceSummary", "MultipleChoiceModelOutput", "None", "True", "__init__", "attention_mask", "attentions", "auto_docstring", "can_return_tuple", "class", "classifier", "conf...
electra/modeling_electra.py:ElectraForCausalLM
[ -0.00029054912738502026, 0.055215027183294296, 0.011065822094678879, -0.001568608800880611, -0.0014331380371004343, 0.02566814236342907, 0.030573610216379166, -0.006588156800717115, 0.0017967700259760022, 0.00958277378231287, 0.02099083736538887, 0.013518555089831352, 0.0005026677972637117, ...
[ "CausalLMOutputWithCrossAttentions", "False", "GenerationMixin", "Linear", "Model", "ModelForCausalLM", "ModelGeneratorPredictions", "ModelModel", "ModelPreTrainedModel", "None", "True", "__init__", "_tied_weights_keys", "add", "attention_mask", "attentions", "auto_docstring", "cac...
wavlm/modeling_wavlm.py:WavLMSamePadLayer
[ -0.00006862095324322581, 0.02618260867893696, 0.03647662326693535, -0.001608440070413053, -0.0004895252641290426, 0.007440784014761448, 0.009734558872878551, -0.04184741526842117, 0.01812642067670822, 0.0049232253804802895, -0.008112132549285889, -0.0022797889541834593, 0.002447626320645213,...
[ "ModelSamePadLayer", "Module", "__init__", "class", "def", "else", "forward", "hidden_states", "if", "nn", "num_conv_pos_embeddings", "num_pad_remove", "return", "self", "super" ]
wavlm/modeling_wavlm.py:WavLMPositionalConvEmbedding
[ -0.00020198767015244812, 0.019476238638162613, 0.01247162651270628, -0.006064968649297953, -0.0009182875510305166, 0.020045718178153038, 0.021640263497829437, -0.03029637038707733, 0.012414677999913692, -0.00808662548661232, 0.020956886932253838, 0.014692600816488266, 0.0030040103010833263, ...
[ "ACT2FN", "Conv1d", "GatheredParameters", "ModelPositionalConvEmbedding", "ModelSamePadLayer", "Module", "__init__", "activation", "class", "config", "conv", "deepspeed", "def", "dim", "else", "feat_extract_activation", "forward", "groups", "hasattr", "hidden_size", "hidden_s...
wavlm/modeling_wavlm.py:WavLMFeatureProjection
[ -0.00018812100461218506, 0.04205431789159775, 0.032105982303619385, 0.02882755547761917, -0.0005016559734940529, 0.021027158945798874, 0.026227423921227455, -0.03775844722986221, 0.0018935747211799026, 0.02532302960753441, 0.01820092648267746, 0.018992271274328232, 0.0018935747211799026, -...
[ "Dropout", "LayerNorm", "Linear", "ModelFeatureProjection", "Module", "__init__", "class", "config", "conv_dim", "def", "dropout", "eps", "feat_proj_dropout", "forward", "hidden_size", "hidden_states", "layer_norm", "layer_norm_eps", "nn", "norm_hidden_states", "projection", ...
wavlm/modeling_wavlm.py:WavLMAttention
[ -0.0002270308177685365, 0.015438094735145569, 0.029059944674372673, -0.03677899017930031, -0.0006633556331507862, 0.034508682787418365, 0.02247605100274086, -0.014075910672545433, 0.0019155724439769983, 0.00683930329978466, 0.03201134502887726, 0.012883998453617096, 0.0004895351594313979, ...
[ "Embedding", "F", "False", "Linear", "ModelAttention", "Module", "None", "Parameter", "True", "ValueError", "_", "__init__", "_relative_positions_bucket", "abs", "add_zero_attn", "and", "arange", "attention_mask", "attn_output", "attn_weights", "be", "bias", "bias_k", "...
wavlm/modeling_wavlm.py:WavLMFeedForward
[ -0.0002094466472044587, 0.0419820211827755, 0.04266650974750519, 0.003279845230281353, -0.0009910836815834045, 0.02372896857559681, 0.04106936603784561, -0.029661210253834724, 0.0015543615445494652, -0.02099101059138775, 0.033768147230148315, -0.008556118234992027, 0.002581095788627863, -0...
[ "ACT2FN", "Dropout", "Linear", "ModelFeedForward", "Module", "__init__", "activation_dropout", "class", "config", "def", "else", "forward", "hidden_act", "hidden_dropout", "hidden_size", "hidden_states", "if", "intermediate_act_fn", "intermediate_dense", "intermediate_dropout",...
wavlm/modeling_wavlm.py:WavLMEncoderLayer
[ -0.00012250700092408806, 0.028315864503383636, 0.04444801062345505, 0.0009236499317921698, -0.0006662970990873873, 0.039709895849227905, 0.025946807116270065, -0.022675253450870514, 0.0062892804853618145, 0.0034689754247665405, 0.002030619652941823, 0.03406928479671478, 0.0027356960345059633...
[ "Dropout", "False", "GradientCheckpointingLayer", "LayerNorm", "ModelAttention", "ModelEncoderLayer", "ModelFeedForward", "None", "True", "__init__", "attention", "attention_dropout", "attention_mask", "attn_residual", "attn_weights", "class", "config", "def", "dropout", "embed...
wavlm/modeling_wavlm.py:WavLMEncoderLayerStableLayerNorm
[ -0.00020959739049430937, 0.0268284659832716, 0.04956445097923279, 0.012447953224182129, -0.0009449644712731242, 0.03751438111066818, 0.017279351130127907, -0.013755272142589092, 0.0062239766120910645, 0.017734069377183914, 0.0033251382410526276, 0.038651179522275925, 0.0022167586721479893, ...
[ "Dropout", "False", "GradientCheckpointingLayer", "LayerNorm", "ModelAttention", "ModelEncoderLayerStableLayerNorm", "ModelFeedForward", "None", "True", "__init__", "attention", "attention_dropout", "attention_mask", "attn_residual", "attn_weights", "class", "config", "def", "dro...
wavlm/modeling_wavlm.py:WavLMEncoder
[ -0.00017923361156135798, 0.03429928049445152, 0.02407764084637165, 0.004344196990132332, -0.0008660000166855752, 0.04474806785583496, 0.007950164377689362, -0.014878164045512676, 0.004060262348502874, -0.004003475420176983, 0.02362334541976452, 0.022260459139943123, 0.002569606527686119, -...
[ "BaseModelOutput", "Dropout", "False", "LayerNorm", "ModelEncoder", "ModelEncoderLayer", "ModelPositionalConvEmbedding", "Module", "ModuleList", "None", "True", "__init__", "all_hidden_states", "all_self_attentions", "and", "attention_mask", "attentions", "class", "config", "de...
wavlm/modeling_wavlm.py:WavLMEncoderStableLayerNorm
[ -0.00014869832375552505, 0.021185971796512604, 0.03308183699846268, 0.024698084220290184, -0.0006372785428538918, 0.03806677088141441, -0.003823671257123351, -0.00827045924961567, 0.0020817765034735203, 0.01229239534586668, 0.016201036050915718, 0.03376160189509392, 0.0016073581064119935, ...
[ "BaseModelOutput", "Dropout", "False", "LayerNorm", "ModelEncoderLayerStableLayerNorm", "ModelEncoderStableLayerNorm", "ModelPositionalConvEmbedding", "Module", "ModuleList", "None", "True", "__init__", "all_hidden_states", "all_self_attentions", "and", "attention_mask", "attentions"...
wavlm/modeling_wavlm.py:WavLMGumbelVectorQuantizer
[ -0.0001172650809166953, 0.026182735338807106, -0.00198909780010581, 0.0033856984227895737, -0.00009963123011402786, 0.0002962486178148538, 0.04965691268444061, -0.005360689479857683, 0, -0.005163190420717001, 0.07177680730819702, 0.002793201245367527, -0.0023982031270861626, 0.007843534462...
[ "FloatTensor", "Linear", "ModelGumbelVectorQuantizer", "Module", "Parameter", "True", "ValueError", "__init__", "_compute_perplexity", "argmax", "batch_size", "be", "by", "class", "codevector_dim", "codevector_idx", "codevector_probs", "codevector_soft_dist", "codevectors", "co...
wavlm/modeling_wavlm.py:WavLMPreTrainedModel
[ -0.0002621118910610676, 0.04610316455364227, 0.00012659825733862817, 0.007474646437913179, -0.0011696965666487813, 0.02556215040385723, 0.030126819387078285, -0.03355032205581665, 0, -0.015405760146677494, 0.027502134442329407, 0.0219104140996933, -0.00018454817472957075, -0.01460694335401...
[ "Conv1d", "False", "GroupNorm", "LayerNorm", "Linear", "Model", "ModelConfig", "ModelFeatureProjection", "ModelGumbelVectorQuantizer", "ModelPositionalConvEmbedding", "ModelPreTrainedModel", "None", "PreTrainedModel", "True", "_conv_out_length", "_get_feat_extract_output_lengths", "_...
wavlm/modeling_wavlm.py:WavLMNoLayerNormConvLayer
[ 0.000045173001126386225, 0.0243696216493845, 0.017374638468027115, 0.00046539207687601447, 0.00048302055802196264, 0.023467043414711952, 0.009871953167021275, -0.03474927693605423, 0.016584880650043488, 0.0016359236324205995, 0.02876969240605831, 0.00823602918535471, 0.004118014592677355, ...
[ "ACT2FN", "Conv1d", "GradientCheckpointingLayer", "ModelNoLayerNormConvLayer", "__init__", "activation", "class", "config", "conv", "conv_dim", "conv_kernel", "conv_stride", "def", "else", "feat_extract_activation", "forward", "hidden_states", "if", "in_conv_dim", "kernel_size"...
wavlm/modeling_wavlm.py:WavLMLayerNormConvLayer
[ -0.000011282235391263384, 0.01938651315867901, 0.025022126734256744, 0.010425886139273643, 0.000192843668628484, 0.03246113657951355, 0.011553009040653706, -0.028741631656885147, 0.016906842589378357, 0.009298763237893581, 0.030883165076375008, 0.015216158702969551, 0.00300096464343369, -0...
[ "ACT2FN", "Conv1d", "GradientCheckpointingLayer", "LayerNorm", "ModelLayerNormConvLayer", "True", "__init__", "activation", "class", "config", "conv", "conv_dim", "conv_kernel", "conv_stride", "def", "elementwise_affine", "else", "feat_extract_activation", "forward", "hidden_st...
wavlm/modeling_wavlm.py:WavLMGroupNormConvLayer
[ -0.000032619624107610434, 0.008350623771548271, -0.005811582785099745, 0.008237777277827263, 0.00018690162687562406, 0.031371261924505234, 0.009761202149093151, -0.023472024127840996, 0.013146590441465378, 0.0025390409864485264, 0.03746496140956879, 0.00668614124879241, 0.001227203174494207,...
[ "ACT2FN", "Conv1d", "GradientCheckpointingLayer", "GroupNorm", "ModelGroupNormConvLayer", "True", "__init__", "activation", "affine", "class", "config", "conv", "conv_dim", "conv_kernel", "conv_stride", "def", "else", "feat_extract_activation", "forward", "hidden_states", "if...
wavlm/modeling_wavlm.py:WavLMFeatureEncoder
[ -0.00006962888437556103, 0.04106517508625984, -0.016358381137251854, 0.0315885990858078, 0, 0.02978353388607502, -0.0009307354339398444, -0.03948574513196945, 0.007671516388654709, 0.007953557185828686, 0.02211201749742031, 0.007333067245781422, 0.003172961762174964, -0.022676100954413414,...
[ "False", "ModelFeatureEncoder", "ModelGroupNormConvLayer", "ModelLayerNormConvLayer", "ModelNoLayerNormConvLayer", "Module", "ModuleList", "None", "True", "ValueError", "__init__", "_freeze_parameters", "_requires_grad", "and", "be", "but", "class", "config", "conv_layer", "con...
wavlm/modeling_wavlm.py:WavLMAdapterLayer
[ -0.00017218169523403049, 0.006077836267650127, 0.032263655215501785, -0.016245432198047638, -0.0007810303941369057, 0.04589618369936943, 0.02010798268020153, -0.03453574329614639, 0.008520331233739853, -0.0059926328249275684, 0.01669984869658947, -0.019994378089904785, 0.0009585372754372656,...
[ "Conv1d", "ModelAdapterLayer", "Module", "__init__", "adapter_kernel_size", "adapter_stride", "class", "config", "conv", "def", "dim", "forward", "functional", "glu", "hidden_states", "nn", "output_hidden_size", "padding", "return", "self", "stride", "super" ]
wavlm/modeling_wavlm.py:WavLMAdapter
[ -0.00026262993924319744, 0.04779507592320442, 0.03750427067279816, 0.00880435574799776, -0.0009218846680596471, 0.0381903238594532, 0.030643733218312263, -0.02504095993936062, -0.005517015233635902, -0.002715629292652011, -0.0015150352846831083, -0.02081029675900936, -0.0023011385928839445, ...
[ "LayerNorm", "Linear", "ModelAdapter", "ModelAdapterLayer", "Module", "ModuleList", "None", "_", "__init__", "and", "class", "config", "def", "else", "for", "forward", "hidden_size", "hidden_states", "if", "in", "is", "layer", "layerdrop", "layerdrop_prob", "layers", ...
wavlm/modeling_wavlm.py:_compute_mask_indices
[ 0, -0.008906038478016853, 0.00551726296544075, -0.014171244576573372, 0, 0.024309566244482994, -0.0006616514874622226, -0.05265205353498459, 0.013331051915884018, -0.0200525913387537, 0.040777336806058884, 0.030695026740431786, -0.006189417093992233, -0.012434846721589565, 0.005489256698...
[ "False", "None", "ValueError", "_", "_compute_mask_indices", "and", "append", "arange", "array", "attention_mask", "batch_size", "be", "bigger", "bool", "broadcast_to", "but", "choice", "compute_num_masked_span", "concatenate", "def", "detach", "dtype", "dummy_mask_idx", ...
wavlm/modeling_wavlm.py:WavLMModel
[ -0.00011478750093374401, 0.03544217348098755, 0.00874838512390852, 0.008075431920588017, -0.0002926643064711243, 0.026132995262742043, 0.022207438945770264, -0.03207740932703018, 0.0029862273950129747, -0.028151853010058403, 0.01351513247936964, 0.011552354320883751, -0.0023132748901844025, ...
[ "False", "ModelAdapter", "ModelBaseModelOutput", "ModelEncoder", "ModelEncoderStableLayerNorm", "ModelFeatureEncoder", "ModelFeatureProjection", "ModelModel", "ModelPreTrainedModel", "None", "Parameter", "Tensor", "True", "__init__", "_compute_mask_indices", "_freeze_parameters", "_g...
wavlm/modeling_wavlm.py:WavLMForCTC
[ -0.0003292037872597575, 0.04213808476924896, 0.0308552086353302, -0.00840459018945694, -0.0011513137724250555, 0.02567429654300213, 0.018305888399481773, -0.053651221096515656, -0.005641437601298094, -0.01726970635354519, 0.03338810056447983, -0.0027775445487350225, -0.00006611059507122263, ...
[ "Cannot", "CausalLMOutput", "Dropout", "False", "Linear", "Model", "ModelForCTC", "ModelModel", "ModelPreTrainedModel", "None", "Please", "True", "ValueError", "You", "_HIDDEN_STATES_START_POSITION", "__class__", "__init__", "_freeze_parameters", "_get_feat_extract_output_lengths...
wavlm/modeling_wavlm.py:WavLMForSequenceClassification
[ -0.0003542772028595209, 0.03962179273366928, 0.0051244948990643024, 0.028055893257260323, -0.0011737669119611382, 0.011852183379232883, 0.006985344458371401, -0.0030059884302318096, -0.011737669818103313, -0.00260519003495574, 0.027712350711226463, -0.0109360720962286, -0.0018035931279882789...
[ "CrossEntropyLoss", "False", "Linear", "Model", "ModelForSequenceClassification", "ModelModel", "ModelPreTrainedModel", "None", "Parameter", "Sequence", "SequenceClassifierOutput", "True", "ValueError", "_HIDDEN_STATES_START_POSITION", "__init__", "_freeze_parameters", "_get_feature_...
wavlm/modeling_wavlm.py:WavLMForAudioFrameClassification
[ -0.00017225838382728398, 0.04364585876464844, 0.0010953866876661777, 0.022840578109025955, -0.0004911572323180735, 0.005879752803593874, 0.023971300572156906, -0.03550466150045395, -0.002770268125459552, -0.016734682023525238, 0.02826804295182228, 0.016395464539527893, -0.0000971713961916975...
[ "Audio", "CrossEntropyLoss", "False", "Linear", "Model", "ModelForAudioFrameClassification", "ModelModel", "ModelPreTrainedModel", "None", "Parameter", "TokenClassifierOutput", "True", "ValueError", "_HIDDEN_STATES_START_POSITION", "__init__", "_freeze_parameters", "adapters", "add...
wavlm/modeling_wavlm.py:AMSoftmaxLoss
[ -0.00015611345588695258, 0.03432026505470276, 0.06367312371730804, 0.007140647619962692, -0.000666789710521698, 0.04628720134496689, 0.028111007064580917, 0.008749409578740597, 0.016144072636961937, 0.012870099395513535, 0.022014643996953964, -0.024837033823132515, 0.002469591563567519, -0...
[ "CrossEntropyLoss", "Model", "Module", "Parameter", "True", "__init__", "bool", "class", "cos_theta", "def", "dim", "flatten", "forward", "functional", "hidden_states", "input_dim", "labels", "logits", "loss", "margin", "mm", "nn", "normalize", "num_labels", "one_hot"...
wavlm/modeling_wavlm.py:TDNNLayer
[ -0.00011568117042770609, 0.028484059497714043, 0.00740359490737319, -0.015372348949313164, -0.00029670895310118794, 0.05448141321539879, 0.03639629855751991, -0.037752680480480194, 0.007460110820829868, -0.008703462779521942, 0.02633645199239254, -0.015598413534462452, 0.001935672713443637, ...
[ "Detected", "Linear", "LoRA", "LoraLayer", "Model", "Module", "ReLU", "You", "__init__", "activation", "applied", "be", "bias", "class", "config", "conv1d", "def", "dilation", "due", "else", "exclude", "forward", "from", "functional", "hidden_states", "if", "in_co...
wavlm/modeling_wavlm.py:WavLMForXVector
[ -0.0002270740078529343, 0.05767679959535599, 0.014646273106336594, 0.02475106716156006, -0.0008124992018565536, 0.04564187675714493, 0.02089080959558487, -0.020323123782873154, -0.006216151174157858, -0.03678598999977112, 0.031336214393377304, 0.0021288187708705664, 0.0021713951136916876, ...
[ "AMSoftmaxLoss", "False", "Linear", "Model", "ModelForXVector", "ModelModel", "ModelPreTrainedModel", "ModuleList", "None", "Parameter", "TDNNLayer", "True", "XVectorOutput", "_HIDDEN_STATES_START_POSITION", "__init__", "_conv_out_length", "_freeze_parameters", "_get_feat_extract_o...
llava_next_video/modeling_llava_next_video.py:LlavaNextVideoModelOutputWithPast
[ -0.00013376322749536484, 0.032666388899087906, 0.004590190481394529, 0.015319408848881721, -0.0008800211944617331, 0.031089389696717262, 0.045057084411382675, -0.05857421085238457, 0.016558479517698288, -0.0008765011443756521, 0.028385963290929794, 0.032666388899087906, -0.00266118417493999,...
[ "BaseModelOutputWithPast", "ModelModelOutputWithPast", "None", "class", "image_hidden_states", "r", "video_hidden_states" ]
llava_next_video/modeling_llava_next_video.py:LlavaNextVideoCausalLMOutputWithPast
[ -0.0002032790653174743, 0.031223664060235023, 0.021155163645744324, 0.007070576306432486, -0.0010040218476206064, 0.032807473093271255, 0.04185781255364418, -0.05022937431931496, 0.017761288210749626, -0.0009121043258346617, 0.02601972036063671, 0.02092890627682209, -0.0012373508652672172, ...
[ "ModelCausalLMOutputWithPast", "ModelOutput", "None", "attentions", "class", "hidden_states", "image_hidden_states", "logits", "loss", "past_key_values", "r", "video_hidden_states" ]
llava_next_video/modeling_llava_next_video.py:LlavaNextVideoPooler
[ -0.0001577583752805367, 0.01455254852771759, 0.004568823147565126, 0.02763856016099453, -0.00028907679370604455, 0.030684443190693855, 0.007501895073801279, -0.02447986789047718, 0.007671110797673464, 0.020757123827934265, 0.02414143644273281, -0.017824051901698112, 0.0031163888052105904, ...
[ "AvgPool2d", "Conv2d", "MaxPool2d", "ModelPooler", "Module", "Unknown", "ValueError", "_", "__init__", "average", "batch_size", "class", "config", "contiguous", "conv", "def", "dim", "elif", "else", "f", "flatten", "forward", "getattr", "hidden_size", "if", "image_f...
llava_next_video/modeling_llava_next_video.py:LlavaNextVideoMultiModalProjector
[ -0.00022233052004594356, 0.04246160015463829, 0.0415581613779068, 0.03433065488934517, -0.0002999697462655604, 0.03229792043566704, 0.022134238854050636, -0.03207205981016159, -0.0032890799921005964, 0.011631768196821213, 0.031620342284440994, -0.005307700019329786, -0.0017433535540476441, ...
[ "ACT2FN", "Linear", "ModelMultiModalProjector", "Module", "__init__", "act", "class", "config", "def", "else", "forward", "hidden_size", "hidden_states", "if", "image_features", "int", "isinstance", "len", "linear_1", "linear_2", "nn", "num_feature_layers", "projector_hid...
llava_next_video/modeling_llava_next_video.py:LlavaNextVideoPreTrainedModel
[ -0.00021475530229508877, 0.0454358346760273, -0.0017748372629284859, 0.017152028158307076, -0.0011287964880466461, 0.029987649992108345, 0.012494854629039764, -0.03044200874865055, -0.0023995800875127316, 0.021014073863625526, 0.019310230389237404, -0.007610502187162638, -0.00352127710357308...
[ "Linear", "LlamaDecoderLayer", "ModelConfig", "ModelModel", "ModelPreTrainedModel", "None", "PreTrainedModel", "True", "_can_compile_fullgraph", "_init_weights", "_no_split_modules", "_skip_keys_device_placement", "_supports_attention_backend", "_supports_flash_attn", "_supports_flex_att...
llava_next_video/modeling_llava_next_video.py:get_anyres_image_grid_shape
[ -0.00020208182104397565, -0.008378488011658192, -0.007647478953003883, 0.008715876378118992, -0.0005763724911957979, -0.0063260393217206, 0.029465286061167717, -0.029465286061167717, 0.010852672159671783, 0.0458848737180233, -0.002853746758773923, -0.0253041572868824, -0.0028396889101713896,...
[ "Model_anyres_image_grid_shape", "Tensor", "TypeError", "a", "be", "def", "either", "f", "grid_pinpoints", "height", "if", "image_size", "invalid", "isinstance", "list", "lists", "ndarray", "not", "np", "of", "or", "patch_size", "raise", "return", "select_best_resolut...
llava_next_video/modeling_llava_next_video.py:image_size_to_num_patches
[ 0.00007069053390296176, -0.009662785567343235, -0.011561829596757889, -0.004580048378556967, 0.0005026882863603532, -0.016756275668740273, 0.026251498609781265, -0.019213862717151642, 0.004971028305590153, 0.00843399204313755, -0.006199821829795837, -0.02424074523150921, -0.00530615402385592...
[ "Model_size", "Model_size_to_num_patches", "Tensor", "TypeError", "a", "be", "best_resolution", "def", "f", "for", "grid_pinpoints", "height", "i", "if", "in", "invalid", "isinstance", "j", "list", "lists", "ndarray", "not", "np", "num_patches", "of", "or", "patch...
llava_next_video/modeling_llava_next_video.py:unpad_image
[ -0.00005201897147344425, 0.028733467683196068, 0.03978480026125908, -0.006022976711392403, -0.00008892870391719043, 0.020444966852664948, -0.0023898507934063673, -0.03801658749580383, 0.007128110155463219, 0.04531046748161316, -0.010554024018347263, -0.016245460137724876, 0.00182346999645233...
[ "Model_image", "Modelded_tensor", "Tensor", "TypeError", "be", "current_aspect_ratio", "current_height", "current_width", "def", "either", "else", "f", "if", "image_size", "int", "invalid", "isinstance", "list", "ndarray", "new_height", "new_width", "not", "np", "or", ...
llava_next_video/modeling_llava_next_video.py:LlavaNextVideoModel
[ -0.00011602733138715848, 0.04287951439619064, -0.011557369492948055, 0.018089795485138893, -0.0002983559970743954, 0.0439961701631546, 0.034839604049921036, -0.05560937151312828, 0.003517460310831666, 0.02947966568171978, 0.026576366275548935, 0.014516502618789673, -0.000016357236745534465, ...
[ "AutoModel", "CLS", "False", "Image", "ModelModel", "ModelModelOutputWithPast", "ModelMultiModalProjector", "ModelPooler", "ModelPreTrainedModel", "None", "Obtains", "Parameter", "True", "ValueError", "Video", "You", "_", "__init__", "_checkpoint_conversion_mapping", "_pixel_va...
llava_next_video/modeling_llava_next_video.py:LlavaNextVideoForConditionalGeneration
[ -0.00021195015870034695, 0.04170338436961174, 0.0005009730812162161, 0.02219696156680584, -0.0005009730812162161, 0.030492795631289482, 0.02578434906899929, -0.05201712250709534, -0.003321136347949505, 0.02769014984369278, 0.03318333625793457, 0.008407940156757832, 0.0016885946970432997, 0...
[ "False", "GenerationMixin", "Linear", "ModelCausalLMOutputWithPast", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "True", "__init__", "_checkpoint_conversion_mapping", "_prepare_4d_causal_attention_mask_with_cache_position", "_tied_weights_keys", "and", ...
sam2_video/modeling_sam2_video.py:Sam2VideoInferenceCache
[ -0.00024386259610764682, 0.02069651335477829, -0.0219405647367239, 0.015720302239060402, -0.0010956146288663149, -0.004834841005504131, 0.022732235491275787, -0.07057171314954758, -0.0012864635791629553, 0.022053660824894905, -0.0017459148075431585, 0.011196473613381386, 0.001809531124308705...
[ "ModelInferenceCache", "None", "Tensor", "True", "__init__", "_vision_features", "and", "cache_vision_features", "cached", "class", "clear", "clear_all", "cpu", "def", "elif", "else", "features", "for", "frame_idx", "get_vision_features", "if", "in", "inference_device", ...
sam2_video/modeling_sam2_video.py:Sam2VideoInferenceSession
[ -0.0000912645336939022, 0.016880009323358536, -0.06125989928841591, 0.01240848284214735, -0.0003336177032906562, 0.019115770235657692, 0.04471525549888611, -0.0733330175280571, 0, 0.0016628485172986984, 0.028505975380539894, 0.02884133905172348, -0.0032977499067783356, -0.00721033476293087...
[ "False", "ModelInferenceCache", "ModelInferenceSession", "None", "OrderedDict", "Tensor", "True", "__init__", "_obj_id_to_idx", "_obj_idx_to_id", "add_mask_inputs", "add_new_frame", "add_point_inputs", "and", "cache", "class", "clear", "clear_all", "cond_frame_outputs", "cpu", ...
sam2_video/modeling_sam2_video.py:Sam2VideoLayerNorm
[ -0.0001214670337503776, 0.036052823066711426, 0.017125090584158897, 0.0576845183968544, -0.000281662680208683, 0.018477071076631546, -0.0048164320178329945, -0.03875678405165672, 0.010590516962110996, 0.047995321452617645, 0.009407533332705498, 0.00543608982115984, 0.003577115945518017, 0....
[ "LayerNorm", "ModelLayerNorm", "NotImplementedError", "Unsupported", "__init__", "channels_first", "channels_last", "class", "data", "data_format", "def", "else", "eps", "f", "features", "format", "forward", "if", "in", "kwargs", "nn", "normalized_shape", "not", "permut...
sam2_video/modeling_sam2_video.py:Sam2VideoPositionEmbeddingSine
[ -0.00023352944117505103, 0.03970357030630112, 0.008956121280789375, 0.015288155525922775, -0.001155168516561389, 0.004535106010735035, 0.042898111045360565, -0.016086790710687637, 0.012093615718185902, 0.016771335154771805, 0.012949296273291111, 0.020194057375192642, -0.0007772429962642491, ...
[ "False", "ModelPositionEmbeddingSine", "Module", "None", "__init__", "and", "arange", "bool", "cat", "class", "compile_compatible_method_lru_cache", "cos", "cumsum", "def", "device", "dim", "dim_t", "div", "dtype", "else", "eps", "flatten", "floor", "forward", "if", ...
sam2_video/modeling_sam2_video.py:eager_attention_forward
[ 0.0000415438917116262, 0.02851148322224617, 0.024890977889299393, -0.012558629736304283, 0.00017766642849892378, 0.03733646497130394, 0.06019090861082077, -0.023080725222826004, 0.022628162056207657, 0.012388918548822403, 0.024438414722681046, 0.02919032797217369, 0.00250324048101902, -0.0...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "softmax", "...
sam2_video/modeling_sam2_video.py:Sam2VideoAttention
[ 0.000022926808014744893, 0.0357210636138916, 0.030104544013738632, -0.008761770091950893, 0.00029486726270988584, 0.035047080367803574, 0.04111292213201523, -0.006795988418161869, 0.00926725659519434, 0.011008378118276596, 0.008537109941244125, 0.020781122148036957, -0.00017814897000789642, ...
[ "ALL_ATTENTION_FUNCTIONS", "Attention", "Falling", "False", "Flash", "Linear", "ModelAttention", "Module", "None", "SDPA", "__init__", "_attn_implementation", "additive", "and", "attention", "attention_downModelple_rate", "attention_interface", "attention_mask", "attention_simila...
sam2_video/modeling_sam2_video.py:Sam2VideoTwoWayAttentionBlock
[ -0.00015365220315288752, 0.02539914660155773, 0.038660649210214615, -0.02472483366727829, -0.00020282089826650918, 0.03731202334165573, 0.022814277559518814, -0.015509214252233505, 0.007642221171408892, 0.020117023959755898, 0.018880782648921013, 0.03506430983543396, -0.00080074742436409, ...
[ "False", "GradientCheckpointingLayer", "LayerNorm", "ModelAttention", "ModelFeedForward", "ModelTwoWayAttentionBlock", "_", "__init__", "attention_similarity", "attn_out", "class", "config", "cross_attn_image_to_token", "cross_attn_token_to_image", "def", "downModelple_rate", "else",...
sam2_video/modeling_sam2_video.py:Sam2VideoFeedForward
[ -0.00016512317233718932, 0.041362464427948, 0.017954036593437195, 0.03136274963617325, -0.0003515525604598224, 0.0420442633330822, 0.046135056763887405, -0.02295389585196972, 0.004346468020230532, -0.014147327281534672, 0.037044405937194824, -0.02977188490331173, 0.003948751837015152, -0.0...
[ "ACT2FN", "F", "False", "Linear", "ModelFeedForward", "Module", "ModuleList", "_", "__init__", "activation", "class", "def", "for", "forward", "hidden_dim", "hidden_states", "if", "in", "input_dim", "layer", "layers", "nn", "num_layers", "output_dim", "proj_in", "pr...
sam2_video/modeling_sam2_video.py:Sam2VideoImageSegmentationOutput
[ -0.00022903038188815117, 0.034088242799043655, -0.001264105667360127, -0.013635297305881977, -0.001015545567497611, 0.03249745815992355, 0.04386020451784134, -0.03181569278240204, 0.01590784639120102, -0.016021473333239555, 0.026475202292203903, 0.030224908143281937, -0.005397304892539978, ...
[ "ModelImageSegmentationOutput", "ModelOutput", "None", "class", "high_res_masks", "image_embeddings", "iou_scores", "mask_decoder_attentions", "object_pointer", "object_score_logits", "pred_masks", "r", "vision_attentions", "vision_hidden_states" ]
sam2_video/modeling_sam2_video.py:Sam2VideoSegmentationOutput
[ -0.000230942721827887, 0.019477959722280502, -0.033456262201070786, -0.008650505915284157, -0.0011672454420477152, 0.045601341873407364, 0.042851511389017105, -0.057288117706775665, 0.01695728302001953, -0.020165417343378067, 0.016269825398921967, 0.04055998846888542, -0.006215760484337807, ...
[ "ModelOutput", "ModelSegmentationOutput", "None", "class", "frame_idx", "object_ids", "object_score_logits", "pred_masks", "r" ]
sam2_video/modeling_sam2_video.py:Sam2VideoPreTrainedModel
[ -0.0002746593963820487, 0.05078147351741791, -0.01631440967321396, 0.011029460467398167, -0.0013930437853559852, 0.03147991746664047, 0.015395288355648518, -0.02964167296886444, -0.005284949671477079, 0.014074050821363926, -0.003877544542774558, 0.010167783126235008, -0.0022978042252361774, ...
[ "Model", "ModelConfig", "ModelMemoryFuserCXBlock", "ModelModel", "ModelPositionalEmbedding", "ModelPreTrainedModel", "ModelVisionRotaryEmbedding", "None", "PreTrainedModel", "True", "_init_weights", "_supports_attention_backend", "_supports_flash_attn", "_supports_sdpa", "base_model_pref...
sam2_video/modeling_sam2_video.py:Sam2VideoVisionRotaryEmbedding
[ -0.00032512337202206254, 0.036067020148038864, 0.021848289296030998, 0.008612156845629215, -0.001444992725737393, 0.03930380195379257, 0.038841404020786285, -0.009479152038693428, 0.0007261088467203081, 0.042309388518333435, -0.0026298868469893932, 0.009768150746822357, -0.000917570374440401...
[ "Dimension", "False", "ModelVisionRotaryEmbedding", "Module", "RoPE", "ValueError", "__init__", "arange", "axial", "be", "by", "cat", "class", "config", "cos", "create_inv_freq", "def", "dim", "div", "divisible", "dtype", "end_x", "end_y", "flattened_indices", "float"...
sam2_video/modeling_sam2_video.py:rotate_pairwise
[ 0.00007118566281860694, 0.010292919352650642, 0.03419724106788635, -0.004190286621451378, 0.000506209151353687, 0.02069832943379879, 0.014961292035877705, -0.014736310578882694, 0.022160710766911507, 0.01889847405254841, 0.015298765152692795, -0.005399564281105995, 0.00099132617469877, -0....
[ "Model_pairwise", "def", "dim", "flatten", "return", "shape", "stack", "start_dim", "torch", "unbind", "view", "x", "x1", "x2" ]
sam2_video/modeling_sam2_video.py:apply_rotary_pos_emb_2d
[ -0.000060345751990098506, 0.017152393236756325, 0.012324732728302479, -0.03362323343753815, -0.00020588550250977278, 0.008803380653262138, 0.04339214414358139, -0.013006284832954407, 0.008462604135274887, 0.021923257037997246, 0.013517448678612709, 0.013858224265277386, 0.0016399845480918884...
[ "False", "Model_rotary_pos_emb_2d", "and", "cat", "cos", "cos_k", "def", "dim", "else", "float", "if", "k", "k_embed", "k_pass", "k_rot", "num_k_exclude_rope", "q", "q_embed", "repeat", "repeat_factor", "repeat_freqs_k", "return", "rotate_pairwise", "shape", "sin", ...
sam2_video/modeling_sam2_video.py:Sam2VideoRoPEAttention
[ -0.0002532076614443213, 0.020025108009576797, 0.019562099128961563, -0.018520331010222435, -0.0008826095145195723, 0.049541886895895004, 0.044217292219400406, -0.006019107531756163, 0.0071476902812719345, 0.02766474522650242, -0.0034580931533128023, 0.013485115952789783, 0.000810264493338763...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "ModelRoPEAttention", "Module", "None", "Tensor", "__init__", "_attn_implementation", "apply_rotary_pos_emb_2d", "attention_interface", "attention_mask", "attn_output", "attn_weights", "batch_size", "class", "config", "contiguous", "c...
sam2_video/modeling_sam2_video.py:Sam2VideoMemoryAttentionLayer
[ -0.00012587856326717883, 0.02677844651043415, 0.023260939866304398, -0.0030494523234665394, -0.0005886152503080666, 0.054464638233184814, 0.02144545130431652, -0.021899323910474777, 0.008793769404292107, 0.04334477335214615, 0.0007304502069018781, 0.033132653683423996, 0.0028225164860486984,...
[ "ACT2FN", "Dropout", "LayerNorm", "Linear", "ModelMemoryAttentionLayer", "ModelRoPEAttention", "Module", "Tensor", "True", "_", "__init__", "activation", "class", "config", "cross_attn_image", "def", "dropout", "dropout1", "dropout2", "dropout3", "forward", "hidden_size", ...
sam2_video/modeling_sam2_video.py:Sam2VideoMemoryAttention
[ -0.0001922139199450612, 0.02494352124631405, 0.009977408684790134, 0.016553428024053574, -0.0007192521006800234, 0.05306167155504227, 0.011904861778020859, -0.02675759419798851, 0.005300498101860285, 0.040816668421030045, 0.005895741283893585, 0.029478706419467926, 0.0030754227191209793, -...
[ "LayerNorm", "ModelMemoryAttention", "ModelMemoryAttentionLayer", "ModelVisionRotaryEmbedding", "Module", "ModuleList", "None", "_", "__init__", "class", "config", "current_vision_features", "current_vision_position_embeddings", "def", "else", "for", "forward", "if", "in", "is"...
sam2_video/modeling_sam2_video.py:Sam2VideoMemoryFuserCXBlock
[ -0.00034702353877946734, 0.013372216373682022, 0.020814144983887672, -0.0076744891703128815, -0.0013735591201111674, 0.021046705543994904, 0.012383834458887577, -0.041628289967775345, 0.005232606083154678, 0.02069786563515663, -0.004302365239709616, -0.011104753240942955, 0.00337212393060326...
[ "ACT2FN", "Conv2d", "GradientCheckpointingLayer", "Linear", "ModelLayerNorm", "ModelMemoryFuserCXBlock", "Parameter", "True", "__init__", "activation", "channels_first", "class", "config", "data_format", "def", "depthwise_conv", "eps", "forward", "groups", "hidden_states", "i...
sam2_video/modeling_sam2_video.py:Sam2VideoMemoryFuser
[ -0.00020246466738171875, 0.011587757617235184, 0.0156406182795763, 0.018152251839637756, -0.0009204561356455088, 0.027856284752488136, 0.009019043296575546, -0.05000431463122368, 0.010788601823151112, 0.01073151919990778, -0.003681824542582035, -0.02705712988972664, 0.0037959895562380552, ...
[ "ModelMemoryFuser", "ModelMemoryFuserCXBlock", "Module", "ModuleList", "_", "__init__", "class", "config", "def", "for", "forward", "hidden_states", "in", "layer", "layers", "memory_fuser_num_layers", "nn", "range", "return", "self", "super" ]
sam2_video/modeling_sam2_video.py:Sam2VideoMaskDownSamplerLayer
[ 0.000046938599552959204, 0.035879600793123245, 0.01624736748635769, -0.00310279568657279, 0.0002997018746100366, 0.052578285336494446, 0.0037515622097998857, -0.02764309011399746, 0.01748848520219326, 0.018165457993745804, 0.009139143861830235, 0.00487985135987401, 0.0038643910083919764, 0...
[ "ACT2FN", "Conv2d", "ModelLayerNorm", "ModelMaskDownModelplerLayer", "Module", "__init__", "activation", "channels_first", "class", "config", "conv", "data_format", "def", "eps", "forward", "in_channels", "kernel_size", "layer_norm", "mask_downModelpler_hidden_act", "mask_downM...