identifier
stringlengths
24
102
embedding
listlengths
2.56k
2.56k
tokens
listlengths
4
448
conditional_detr/modeling_conditional_detr.py:ConditionalDetrPreTrainedModel
[ -0.00016010223771445453, 0.05089491605758667, -0.00605221651494503, 0.015426114201545715, -0.0007811581599526107, 0.029275838285684586, 0.021731680259108543, -0.011147337965667248, 0.003659479785710573, 0.010471741668879986, 0.007938255555927753, -0.004503975156694651, -0.001731215394102037,...
[ "Conv2d", "Embedding", "False", "GroupNorm", "LayerNorm", "Linear", "ModelConfig", "ModelConvEncoder", "ModelDecoderLayer", "ModelEncoderLayer", "ModelLearnedPositionEmbedding", "ModelMHAttentionMap", "ModelPreTrainedModel", "None", "PreTrainedModel", "_init_weights", "_is_hf_initial...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrEncoder
[ 0, 0.03743509203195572, 0.01668793335556984, 0.021536454558372498, 0, 0.04374944791197777, 0.014376428909599781, -0.013079731725156307, 0.0062297857366502285, -0.010260824114084244, 0.017364470288157463, 0.025595681741833687, -0.0023678825236856937, -0.005384113639593124, -0.021085429936...
[ "BaseModelOutput", "False", "ModelEncoder", "ModelEncoderLayer", "ModelPreTrainedModel", "ModuleList", "None", "True", "_", "__init__", "_prepare_4d_attention_mask", "all_attentions", "attention_mask", "attentions", "class", "config", "def", "dropout", "dropout_probability", "d...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrDecoder
[ -0.00017772398132365197, 0.06881189346313477, -0.005291046109050512, 0.0024474626407027245, -0.0008099261904135346, 0.033274173736572266, 0.05183527618646622, -0.024785863235592842, 0.004753453191369772, -0.007696067448705435, -0.011204568669199944, 0.02738894522190094, -0.004696864634752273...
[ "False", "LayerNorm", "MLP", "ModelDecoder", "ModelDecoderLayer", "ModelDecoderOutput", "ModelPreTrainedModel", "ModuleList", "None", "_", "__init__", "_prepare_4d_attention_mask", "all_cross_attentions", "all_hidden_states", "all_self_attns", "and", "attention_mask", "attentions",...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrModel
[ -0.00003559171091183089, 0.028172913938760757, 0.006791684776544571, 0.012968484312295914, -0.00010131628368981183, 0.03756388649344444, 0.013415672816336155, -0.015539821237325668, 0.01045304536819458, -0.0010061755310744047, 0.005925255827605724, 0.01268899068236351, -0.0007022266509011388...
[ "BaseModelOutput", "Conv2d", "Embedding", "False", "ModelConvEncoder", "ModelConvModel", "ModelDecoder", "ModelEncoder", "ModelModel", "ModelModelOutput", "ModelPreTrainedModel", "None", "True", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "backbone", "b...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrMLPPredictionHead
[ -0.00024052357184700668, -0.0006734660128131509, 0.02770831435918808, 0.04241766780614853, -0.0006734660128131509, 0.040137145668268204, 0.04583844542503357, -0.009635196067392826, -0.0012685391120612621, -0.013569092378020287, 0.022007014602422714, -0.0588374100625515, -0.002750876825302839...
[ "Linear", "ModelMLPPredictionHead", "Module", "ModuleList", "__init__", "class", "def", "else", "enumerate", "for", "forward", "functional", "h", "hidden_dim", "i", "if", "in", "input_dim", "k", "layer", "layers", "n", "nn", "num_layers", "output_dim", "relu", "re...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrForObjectDetection
[ -0.000123675970826298, 0.014146425761282444, -0.0028770011849701405, 0.007241622544825077, -0.00038418491021730006, 0.03300832584500313, 0.03615197539329529, -0.01773916743695736, 0.00611889036372304, 0.007017076015472412, 0.005754002369940281, 0.04378655552864075, -0.003045411081984639, -...
[ "Linear", "ModelForObjectDetection", "ModelMLPPredictionHead", "ModelModel", "ModelObjectDetectionOutput", "ModelPreTrainedModel", "None", "__init__", "_set_aux_loss", "a", "append", "auto_docstring", "auxiliary_loss", "auxiliary_outputs", "b", "bbox_predictor", "class", "class_lab...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrForSegmentation
[ -0.00018483103485777974, 0.026136867702007294, -0.021292533725500107, -0.0051541454158723354, -0.0004928827402181923, 0.038078710436820984, 0.023658372461795807, -0.03019258752465248, 0.004872498102486134, -0.002957296557724476, 0.014082363806664944, 0.033797673881053925, -0.0035487557761371...
[ "BaseModelOutput", "Model", "ModelForObjectDetection", "ModelForSegmentation", "ModelMHAttentionMap", "ModelMaskHeadSmallConv", "ModelPreTrainedModel", "ModelSegmentationOutput", "None", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "auxiliary_loss", "auxiliary_o...
conditional_detr/modeling_conditional_detr.py:_expand
[ -0.00022749749768991023, -0.030248349532485008, 0.04176078364253044, 0.006292334292083979, -0.000828866905067116, 0.046049728989601135, 0.026072273030877113, -0.06410844624042511, 0.003244925756007433, 0.013600470498204231, 0.0013755663530901074, -0.02934541366994381, -0.0020033889450132847,...
[ "_expand", "def", "flatten", "int", "length", "repeat", "return", "tensor", "unsqueeze" ]
conditional_detr/modeling_conditional_detr.py:ConditionalDetrMaskHeadSmallConv
[ -0.00021306339476723224, -0.00419351551681757, 0.013065510429441929, -0.013521947897970676, -0.0007559738587588072, 0.03354812413454056, 0.027043895795941353, -0.05773928761482239, 0.008843468502163887, 0.012209691107273102, 0.00456437049433589, 0.007360047660768032, 0.0015048158820718527, ...
[ "Conv2d", "GroupNorm", "ModelMaskHeadSmallConv", "Module", "The", "ValueError", "__init__", "_expand", "a", "adapter1", "adapter2", "adapter3", "as", "attention", "bbox_mask", "be", "bias", "by", "cat", "class", "constant_", "context_dim", "cur_fpn", "def", "dim", "...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrMHAttentionMap
[ -0.00014680984895676374, 0.02060290426015854, 0.051394056528806686, -0.03667769581079483, -0.00048464935389347374, 0.019244471564888954, 0.04641313478350639, -0.022866958752274513, 0.008886417374014854, 0.022187743335962296, 0.01550877932459116, 0.030338343232870102, 0.000502337294165045, ...
[ "Dropout", "Linear", "ModelMHAttentionMap", "Module", "None", "__init__", "bias", "bnchw", "bqnc", "bqnhw", "class", "conv2d", "def", "dim", "dropout", "dtype", "einsum", "finfo", "flatten", "float", "forward", "functional", "hidden_dim", "if", "is", "k", "k_linea...
flaubert/modeling_flaubert.py:create_sinusoidal_embeddings
[ -0.00018755829660221934, 0.03315316140651703, 0.028237348422408104, 0.03452501818537712, -0.0008931347401812673, 0.024807710200548172, 0.05578877031803131, -0.010917679406702518, 0.01737682893872261, 0.016576580703258514, 0.018177079036831856, 0.004744331818073988, 0.0004501399234868586, -...
[ "False", "FloatTensor", "Model_sinusoidal_embeddings", "array", "cos", "def", "detach_", "dim", "for", "in", "j", "n_pos", "np", "out", "pos", "position_enc", "power", "range", "requires_grad", "return", "sin", "torch" ]
flaubert/modeling_flaubert.py:get_masks
[ -0.00004481861833482981, -0.003693928709253669, 0.033357296139001846, -0.011865346692502499, 0.00019151761080138385, 0.04119290038943291, -0.0020568466279655695, -0.05731186270713806, 0.019141266122460365, 0.02294713258743286, 0.023618755862116814, 0.018917392939329147, 0.0024906033650040627...
[ "False", "Model_masks", "None", "alen", "arange", "assert", "attn_mask", "bs", "causal", "def", "device", "dtype", "else", "if", "is", "item", "lengths", "long", "mask", "not", "or", "padding_mask", "repeat", "return", "size", "slen", "torch" ]
flaubert/modeling_flaubert.py:MultiHeadAttention
[ -0.00016036011220421642, 0.01905994303524494, 0.03766876459121704, -0.0046240100637078285, -0.000874050660058856, 0.01488705724477768, 0.038796573877334595, -0.024586200714111328, 0.0009092946420423687, 0.03157860413193703, -0.014097591862082481, 0.019172724336385727, -0.0013815639540553093,...
[ "EncoderDecoderCache", "False", "Linear", "ModelHeadAttention", "Module", "None", "True", "__init__", "and", "assert", "attention_dropout", "bs", "cache", "cache_position", "class", "config", "context", "contiguous", "cross_attention_cache", "curr_past_key_values", "current_s...
flaubert/modeling_flaubert.py:TransformerFFN
[ -0.000245455012191087, 0.03669283911585808, 0.02373567968606949, 0.020295726135373116, -0.000670074310619384, 0.017543762922286987, 0.042655423283576965, -0.03715150058269501, 0.004701269790530205, 0.0041566104628145695, 0.014791800640523434, -0.017085103318095207, 0.003841281635686755, -0...
[ "Linear", "ModelFFN", "Module", "__init__", "act", "apply_chunking_to_forward", "chunk_size_feed_forward", "class", "config", "def", "dim_hidden", "dropout", "else", "ff_chunk", "forward", "functional", "gelu", "gelu_activation", "if", "in_dim", "input", "lin1", "lin2", ...
flaubert/modeling_flaubert.py:FlaubertPredLayer
[ -0.00039676710730418563, 0.03571266308426857, 0.017740381881594658, 0.028871601447463036, -0.0015870684292167425, 0.03385746106505394, 0.03779976814985275, -0.007246887777000666, -0.002971224021166563, -0.015305427834391594, 0.024581443518400192, 0.0012247241102159023, -0.0008370155701413751...
[ "AdaptiveLogSoftmaxWithLoss", "False", "Linear", "ModelPredLayer", "Module", "None", "True", "__init__", "asm", "asm_cutoffs", "asm_div_value", "class", "config", "cross_entropy", "cutoffs", "def", "dim", "div_value", "else", "emb_dim", "forward", "functional", "head_bias...
flaubert/modeling_flaubert.py:FlaubertSquadHeadOutput
[ -0.0004275301762390882, 0.014213989488780499, 0.01658298820257187, 0.016701437532901764, -0.001969229895621538, 0.05993565917015076, 0.077229343354702, -0.002872410463169217, 0.008587619289755821, 0.00458993436768651, 0.023453082889318466, 0.03198147565126419, -0.002990860491991043, -0.004...
[ "ModelOutput", "ModelSquadHeadOutput", "None", "class", "cls_logits", "end_top_index", "end_top_log_probs", "loss", "r", "start_top_index", "start_top_log_probs" ]
flaubert/modeling_flaubert.py:FlaubertPoolerStartLogits
[ -0.00026344796060584486, 0.01514019351452589, 0.02305438555777073, 0.027068831026554108, -0.0012903574388474226, 0.04450299218297005, 0.0399150550365448, -0.020531021058559418, 0.0017706571379676461, 0.006652509327977896, 0.019728131592273712, 0.0014982483116909862, 0.00007079044007696211, ...
[ "Linear", "ModelPoolerStartLogits", "Module", "None", "__init__", "class", "config", "def", "dense", "else", "float16", "forward", "hidden_size", "hidden_states", "if", "is", "nn", "not", "p_mask", "return", "self", "squeeze", "super", "torch", "x" ]
flaubert/modeling_flaubert.py:FlaubertPoolerEndLogits
[ -0.0003396792453713715, 0.021855106577277184, 0.03885352239012718, 0.03538445755839348, -0.0014526707818731666, 0.05550503358244896, 0.040934961289167404, -0.029255777597427368, 0.004394148476421833, 0.017229687422513962, 0.03145285323262215, 0.021161293610930443, 0.0020669843070209026, -0...
[ "LayerNorm", "Linear", "ModelPoolerEndLogits", "Module", "None", "One", "Tanh", "__init__", "activation", "assert", "be", "cat", "class", "config", "def", "dense_0", "dense_1", "dim", "else", "eps", "expand", "float16", "forward", "gather", "hidden_size", "hidden_st...
flaubert/modeling_flaubert.py:FlaubertPoolerAnswerClass
[ -0.0003486787900328636, 0.016564039513468742, 0.02576628513634205, 0.03427835926413536, -0.0012437409022822976, 0.046931445598602295, 0.030597463250160217, 0.0007620608666911721, -0.0016391498502343893, 0.01851951703429222, 0.04509099945425987, 0.017484264448285103, 0.0027175378054380417, ...
[ "Linear", "ModelPoolerAnswerClass", "Module", "None", "One", "Tanh", "__init__", "activation", "assert", "be", "cat", "class", "cls_index", "cls_token_state", "config", "def", "dense_0", "dense_1", "dim", "else", "expand", "forward", "gather", "hidden_size", "hidden_s...
flaubert/modeling_flaubert.py:FlaubertSQuADHead
[ -0.00046064614434726536, 0.025353962555527687, 0.027240769937634468, 0.026533218100667, -0.001931028557009995, 0.052594732493162155, 0.04481165483593941, 0.0023290268145501614, -0.002137398114427924, 0.019103916361927986, 0.027830397710204124, 0.026533218100667, 0.0014224753249436617, -0.0...
[ "BCEWithLogitsLoss", "CrossEntropyLoss", "False", "ModelPoolerAnswerClass", "ModelPoolerEndLogits", "ModelPoolerStartLogits", "ModelSQuADHead", "ModelSquadHeadOutput", "Module", "None", "__init__", "and", "answer_class", "auto_docstring", "bh", "bl", "blh", "class", "cls_index", ...
flaubert/modeling_flaubert.py:FlaubertSequenceSummary
[ -0.00037931749830022454, 0.022525623440742493, 0.020658213645219803, 0.04575152322649956, -0.0016193939372897148, 0.025326736271381378, 0.020074648782610893, -0.01207980327308178, -0.0034576246980577707, 0.00682771485298872, 0.02917826734483242, 0.020658213645219803, 0.0016704559093341231, ...
[ "Dropout", "Identity", "Linear", "ModelSequenceSummary", "Module", "None", "NotImplementedError", "__init__", "activation", "activation_string", "and", "attn", "class", "cls_index", "config", "def", "dim", "dtype", "elif", "else", "expand", "first", "first_dropout", "fo...
flaubert/modeling_flaubert.py:FlaubertPreTrainedModel
[ -0.0002607954083941877, 0.05853249132633209, 0.003272446570917964, 0.02663685753941536, -0.0012646791292354465, 0.0281230341643095, 0.027094142511487007, -0.015662014484405518, -0.004572851117700338, -0.0046585919335484505, 0.005973286461085081, 0, -0.0015076118288561702, -0.00194346159696...
[ "Embedding", "False", "LayerNorm", "Linear", "ModelConfig", "ModelModel", "ModelPreTrainedModel", "None", "PreTrainedModel", "_init_weights", "_is_hf_initialized", "and", "arange", "attention_mask", "attns_list", "base_model_prefix", "bias", "class", "config", "copy_", "creat...
flaubert/modeling_flaubert.py:FlaubertModel
[ -0.00005615027475869283, 0.05637487769126892, 0.01605897955596447, 0.008478691801428795, -0.000096069612482097, 0.05300585925579071, 0.021000202745199203, -0.02212320826947689, 0.006092304829508066, -0.0046323975548148155, 0.017406584694981575, 0.0099947489798069, 0.00006316906365100294, 0...
[ "BaseModelOutput", "DynamicCache", "Embedding", "EncoderDecoderCache", "False", "FloatTensor", "LayerNorm", "ModelModel", "ModelPreTrainedModel", "ModuleList", "MultiHeadAttention", "None", "TransformerFFN", "__init__", "_slen", "a", "and", "append", "arange", "assert", "atte...
flaubert/modeling_flaubert.py:FlaubertWithLMHeadModel
[ -0.00022195327619556338, 0.022434895858168602, 0.006595182698220015, -0.01713620126247406, -0.0007539365324191749, 0.0446443147957325, 0.02390049211680889, -0.01454322412610054, -0.003114391816779971, -0.00873720832169056, 0.013923163525760174, 0.016234295442700386, -0.0020574715454131365, ...
[ "GenerationMixin", "MaskedLMOutput", "ModelModel", "ModelPreTrainedModel", "ModelPredLayer", "ModelWithLMHeadModel", "None", "Tensor", "__init__", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "cache", "cat", "class", "config", "def", "device", "dim", ...
flaubert/modeling_flaubert.py:FlaubertForSequenceClassification
[ -0.0003979328612331301, 0.028233427554368973, 0.014981001615524292, 0.02131911739706993, -0.0012820280389860272, 0.031805820763111115, 0.013770997524261475, -0.001447683316655457, -0.006251687183976173, 0.002477627247571945, 0.046556346118450165, 0.0063669257797300816, -0.0002187730424338951...
[ "BCEWithLogitsLoss", "CrossEntropyLoss", "MSELoss", "ModelForSequenceClassification", "ModelModel", "ModelPreTrainedModel", "ModelSequenceSummary", "None", "SequenceClassifierOutput", "Tensor", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "cache", "class", "...
flaubert/modeling_flaubert.py:FlaubertForTokenClassification
[ -0.00028927106177434325, 0.03725525364279747, 0.009828073903918266, -0.026970062404870987, -0.0010928018018603325, 0.04251212999224663, 0.04182645305991173, 0.0007035357994027436, -0.00214274856261909, 0.004942606668919325, 0.05348300561308861, 0.02331310510635376, -0.0013927866239100695, ...
[ "CrossEntropyLoss", "Dropout", "Linear", "ModelForTokenClassification", "ModelModel", "ModelPreTrainedModel", "None", "Tensor", "TokenClassifierOutput", "__init__", "attention_mask", "attentions", "auto_docstring", "cache", "class", "classifier", "config", "def", "dropout", "el...
flaubert/modeling_flaubert.py:FlaubertForQuestionAnsweringSimple
[ -0.00017516175284981728, 0.030870718881487846, 0.02557537704706192, 0.025800710543990135, -0.0005598134594038129, 0.04979874938726425, 0.03740539774298668, 0.017238028347492218, 0.0004401049227453768, 0.02827938087284565, 0.017238028347492218, 0.018928032368421555, 0.0012322937836870551, 0...
[ "CrossEntropyLoss", "Linear", "ModelForQuestionAnsweringSimple", "ModelModel", "ModelPreTrainedModel", "None", "QuestionAnsweringModelOutput", "Tensor", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "cache", "clamp", "class", "config", "contiguous", "def", ...
flaubert/modeling_flaubert.py:FlaubertForQuestionAnsweringOutput
[ -0.0002433365589240566, 0.02347482182085514, 0.029314899817109108, 0.025077980011701584, -0.0012166828382760286, 0.06229415908455849, 0.07237115502357483, 0.013283313252031803, 0.012023689225316048, 0.004523197188973427, 0.016489630565047264, 0.029200388118624687, -0.001116485451348126, -0...
[ "ModelForQuestionAnsweringOutput", "ModelOutput", "None", "attentions", "class", "cls_logits", "end_top_index", "end_top_log_probs", "hidden_states", "loss", "r", "start_top_index", "start_top_log_probs" ]
flaubert/modeling_flaubert.py:FlaubertForQuestionAnswering
[ -0.0002513910294510424, 0.030138598755002022, 0.016995450481772423, 0.01858169212937355, -0.000810824625659734, 0.047587260603904724, 0.04985332116484642, 0.008271118625998497, 0.007364694960415363, 0.01994132809340954, 0.01812848076224327, 0.01960141956806183, 0.0018128480296581984, -0.00...
[ "ModelForQuestionAnswering", "ModelForQuestionAnsweringOutput", "ModelModel", "ModelPreTrainedModel", "ModelSQuADHead", "None", "Tensor", "__init__", "attention_mask", "attentions", "auto_docstring", "cache", "class", "cls_index", "cls_logits", "config", "def", "else", "end_posit...
flaubert/modeling_flaubert.py:FlaubertForMultipleChoice
[ -0.00026341210468672216, 0.049208227545022964, 0.025629285722970963, 0.03508364409208298, -0.0009183827205561101, 0.04237375408411026, 0.04009559378027916, -0.01879481039941311, -0.004072208888828754, -0.00025451305555179715, 0.030527327209711075, -0.00023671492817811668, -0.0034599534701555...
[ "CrossEntropyLoss", "Linear", "Model", "ModelForMultipleChoice", "ModelModel", "ModelPreTrainedModel", "ModelSequenceSummary", "MultipleChoiceModelOutput", "None", "Please", "Tensor", "The", "__init__", "attention", "attention_mask", "attentions", "auto_docstring", "be", "cache",...
regnet/modeling_regnet.py:RegNetConvLayer
[ -0.000010332522833778057, 0.0037259298842400312, 0.005830653943121433, -0.002830000128597021, 0.00026486811111681163, 0.036406032741069794, 0.0023749247193336487, -0.016837790608406067, 0.02036462537944317, 0.0006968342349864542, 0.04505246505141258, -0.03026251494884491, 0.00308598019182682...
[ "ACT2FN", "BatchNorm2d", "Conv2d", "Identity", "ModelConvLayer", "Module", "None", "__init__", "activation", "class", "convolution", "def", "else", "forward", "groups", "hidden_state", "if", "in_channels", "is", "kernel_size", "nn", "normalization", "not", "out_channels...
regnet/modeling_regnet.py:RegNetEmbeddings
[ -0.000047152152546914294, 0.013650327920913696, 0.01737314462661743, 0.01455282885581255, 0, 0.03203878551721573, 0.019629396498203278, -0.03497191518545151, 0.01235298253595829, 0.009532666765153408, 0.026059716939926147, -0.008912197314202785, 0.0008672470576129854, -0.002369065303355455...
[ "Make", "ModelConvLayer", "ModelEmbeddings", "Module", "ValueError", "__init__", "activation", "channel", "class", "config", "configuration", "def", "dimension", "embedder", "embedding_size", "forward", "hidden_act", "hidden_state", "if", "in", "kernel_size", "match", "nn...
regnet/modeling_regnet.py:RegNetShortCut
[ -0.00010893669241340831, -0.002579054096713662, 0.029928363859653473, 0.006121710874140263, -0.00018598948372527957, 0.031062014400959015, 0.016211196780204773, -0.01995224319398403, 0.014000579714775085, 0.014397356659173965, 0.02448684349656105, -0.031288743019104004, 0.002933319890871644,...
[ "BatchNorm2d", "Conv2d", "ModelShortCut", "Module", "__init__", "class", "convolution", "def", "forward", "hidden_state", "in_channels", "input", "kernel_size", "nn", "normalization", "out_channels", "return", "self", "stride", "super" ]
regnet/modeling_regnet.py:RegNetSELayer
[ -0.00002075294469250366, 0.042728107422590256, 0.010060321539640427, 0.0035324161872267723, 0.00026669743238016963, 0.03888483718037605, 0.022042278200387955, -0.01639041118323803, 0.012999291531741619, 0.009382097981870174, 0.024189986288547516, -0.0012363457353785634, 0.002274875994771719,...
[ "AdaptiveAvgPool2d", "Conv2d", "ModelSELayer", "Module", "ReLU", "Sequential", "Sigmoid", "__init__", "attention", "class", "def", "forward", "hidden_state", "in_channels", "kernel_size", "nn", "pooled", "pooler", "reduced_channels", "return", "self", "super" ]
regnet/modeling_regnet.py:RegNetXLayer
[ -0.00020149024203419685, -0.012553020380437374, 0.017916584387421608, 0.0011982428841292858, -0.0006668792339041829, 0.058428604155778885, 0.0057915071956813335, -0.036517877131700516, 0.006704454310238361, -0.023964857682585716, 0.028529591858386993, -0.03628964349627495, 0.0021967785432934...
[ "ACT2FN", "Identity", "ModelConvLayer", "ModelShortCut", "ModelXLayer", "Module", "None", "Sequential", "__init__", "activation", "class", "config", "def", "else", "forward", "groups", "groups_width", "hidden_act", "hidden_state", "if", "in_channels", "kernel_size", "laye...
regnet/modeling_regnet.py:RegNetYLayer
[ -0.00017524529539514333, -0.0038235336542129517, 0.016993483528494835, 0.004418305587023497, -0.00018763638217933476, 0.05347282811999321, 0.025830093771219254, -0.03783882409334183, 0.00594771932810545, 0, 0.03851856291294098, -0.02684970386326313, 0.0006939005688764155, -0.01189543865621...
[ "ACT2FN", "Identity", "ModelConvLayer", "ModelSELayer", "ModelShortCut", "ModelYLayer", "Module", "None", "Sequential", "__init__", "activation", "class", "config", "def", "else", "forward", "groups", "groups_width", "hidden_act", "hidden_state", "if", "in_channels", "int...
regnet/modeling_regnet.py:RegNetStage
[ -0.00008603272726759315, -0.014304493553936481, -0.0008053373312577605, 0.00516551174223423, 0.00009179781045531854, 0.062213193625211716, 0.022137906402349472, -0.04086998105049133, 0.007322538178414106, 0.011636591516435146, 0.04904397949576378, -0.05562858656048775, 0.0007272869697771966,...
[ "ModelStage", "ModelXLayer", "ModelYLayer", "Module", "Sequential", "_", "__init__", "class", "config", "def", "depth", "else", "for", "forward", "hidden_state", "if", "in", "in_channels", "layer", "layer_type", "layers", "nn", "out_channels", "range", "return", "se...
regnet/modeling_regnet.py:RegNetEncoder
[ -0.00009247014531865716, 0.0033958558924496174, 0.01228708028793335, 0.024235984310507774, 0, 0.036973968148231506, 0.011103462427854538, -0.04463930055499077, 0.0060308147221803665, 0.007778060156852007, 0.03584671393036842, -0.011441639624536037, -0.000979302916675806, 0.0056080939248204...
[ "BaseModelOutputWithNoAttention", "False", "ModelEncoder", "ModelStage", "Module", "ModuleList", "None", "True", "__init__", "append", "class", "config", "def", "depth", "depths", "downsample_in_first_stage", "else", "embedding_size", "for", "forward", "hidden_sizes", "hidd...
regnet/modeling_regnet.py:RegNetPreTrainedModel
[ -0.000033521548175485805, 0.041505370289087296, -0.008637604303658009, 0.01592908799648285, 0.00019192729087080806, 0.01985527202486992, 0.018621329218149185, -0.010264166630804539, 0.007908456027507782, 0.014302526600658894, 0.01873350515961647, -0.02265968918800354, -0.0009464907343499362,...
[ "BatchNorm2d", "Conv2d", "GroupNorm", "Linear", "Model", "ModelConfig", "ModelPreTrainedModel", "ModelYLayer", "None", "PreTrainedModel", "_", "_calculate_fan_in_and_fan_out", "_init_weights", "_no_split_modules", "a", "base_model_prefix", "bias", "bound", "class", "config", ...
regnet/modeling_regnet.py:RegNetModel
[ 0.00004328732393332757, 0.025745024904608727, 0.010969619266688824, 0.03716238588094711, 0.0003707843425218016, 0.028655333444476128, 0.022386979311704636, -0.01555895060300827, 0.01304041501134634, 0.030222421512007713, 0.026752440258860588, -0.003008250379934907, 0.0018749095033854246, -...
[ "AdaptiveAvgPool2d", "BaseModelOutputWithPoolingAndNoAttention", "ModelEmbeddings", "ModelEncoder", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "auto_docstring", "class", "config", "def", "else", "embedder", "embedding_output", "encoder", "encoder_outputs", "forward",...
regnet/modeling_regnet.py:RegNetForImageClassification
[ -0.000042190989915980026, 0.031177837401628494, 0.008740929886698723, 0.027837354689836502, 0.00009003644663607702, 0.006207730155438185, 0.02171313762664795, -0.008295531384646893, 0.0046488381922245026, 0.00823985692113638, 0.047212153673172, 0.009631725028157234, 0.000953429436776787, -...
[ "Flatten", "Identity", "ImageClassifierOutputWithNoAttention", "Linear", "Model", "ModelForImageClassification", "ModelModel", "ModelPreTrainedModel", "None", "Sequential", "__init__", "auto_docstring", "class", "classifier", "config", "def", "else", "forward", "hidden_sizes", ...
glm4_moe/modeling_glm4_moe.py:Glm4MoeRotaryEmbedding
[ -0.00030098692514002323, 0.04780977591872215, 0.004612714983522892, -0.006150286644697189, -0.0016971309669315815, 0.03643754869699478, 0.03875841200351715, 0.012590681202709675, -0.004351617768406868, 0.02146798186004162, 0.002509433077648282, -0.0033362405374646187, -0.0013707596808671951,...
[ "False", "ModelRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "arange", "attention_factor", "attention_scaling", "base", "cat", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cpu", "def", "default", "device", ...
glm4_moe/modeling_glm4_moe.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
glm4_moe/modeling_glm4_moe.py:eager_attention_forward
[ 0, 0.021594731137156487, 0.01775064319372177, -0.018768195062875748, -0.00008832923776935786, 0.039797618985176086, 0.05359111353754997, -0.035049039870500565, 0.02069023996591568, 0.011645326390862465, 0.029395969584584236, 0.022499222308397293, 0.002741739386692643, -0.014584923163056374...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "causal_mask", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "key_states", "kwargs", "matmul", "module", "nn", "not", "num_key_value_groups", ...
glm4_moe/modeling_glm4_moe.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
glm4_moe/modeling_glm4_moe.py:apply_rotary_pos_emb
[ -0.00020313466666266322, 0.028396088629961014, 0.021097494289278984, -0.006728390231728554, -0.0008695589494891465, 0.02337830513715744, 0.04196690768003464, -0.0029650533106178045, 0.011575112119317055, 0.03329982981085777, 0.0030790939927101135, 0.00803985632956028, -0.0009693444008007646,...
[ "Model_rotary_pos_emb", "cat", "cos", "def", "dim", "k", "k_embed", "k_pass", "k_rot", "q", "q_embed", "q_pass", "q_rot", "return", "rotary_dim", "rotate_half", "shape", "sin", "torch", "unsqueeze", "unsqueeze_dim" ]
glm4_moe/modeling_glm4_moe.py:Glm4MoeAttention
[ -0.000058884113968815655, 0.038698285818099976, 0.019349142909049988, 0.002812375547364354, -0.00020126062736380845, 0.0247489046305418, 0.035098448395729065, -0.013836887665092945, 0.00023729418171569705, 0.02441141940653324, 0.018224192783236504, 0.03127361461520195, 0.0004499800852499902,...
[ "ALL_ATTENTION_FUNCTIONS", "Linear", "ModelAttention", "ModelRMSNorm", "Module", "None", "Tensor", "True", "__init__", "_attn_implementation", "apply_rotary_pos_emb", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "cache_kwargs", "cache...
glm4_moe/modeling_glm4_moe.py:Glm4MoeMLP
[ -0.00024263472005259246, 0.02898676134645939, 0.02358049899339676, 0.02829659916460514, -0.0008483228739351034, 0.05774346739053726, 0.029906975105404854, -0.005061180330812931, -0.001998591236770153, -0.004831126891076565, 0.028066545724868774, -0.04946153610944748, -0.0024299416691064835, ...
[ "ACT2FN", "Linear", "ModelMLP", "Module", "None", "__init__", "act_fn", "class", "config", "def", "down_proj", "else", "forward", "gate_proj", "hidden_act", "hidden_size", "if", "intermediate_size", "is", "nn", "return", "self", "super", "up_proj", "x" ]
glm4_moe/modeling_glm4_moe.py:Glm4MoeTopkRouter
[ -0.00039330136496573687, 0.03474346548318863, -0.0025112200528383255, -0.00980853009968996, -0.001624907017685473, 0.06617803126573563, 0.07326853275299072, -0.030134640634059906, -0.008390429429709911, 0.010576668195426464, 0.021980561316013336, -0.020916985347867012, -0.0009823301807045937...
[ "F", "ModelTopkRouter", "Module", "Parameter", "__init__", "class", "config", "def", "dtype", "e_score_correction_bias", "empty", "float32", "forward", "hidden_size", "hidden_states", "linear", "n_group", "n_routed_experts", "nn", "norm_topk_prob", "num_experts_per_tok", "r...
glm4_moe/modeling_glm4_moe.py:Glm4MoeRMSNorm
[ -0.00009884802420856431, 0.04179859161376953, 0.03230918198823929, 0.053095508366823196, -0.00046952810953371227, 0.03931327164173126, 0.022028988227248192, -0.02937198430299759, 0.007964326068758965, 0.042476408183574677, 0.019995542243123055, 0.006975846365094185, 0.0028524715453386307, ...
[ "ModelRMSNorm", "Module", "Parameter", "True", "__init__", "class", "def", "eps", "extra_repr", "f", "float32", "forward", "hidden_size", "hidden_states", "keepdim", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "shape", "super", "to", "torch", "tuple", ...
glm4_moe/modeling_glm4_moe.py:Glm4MoeNaiveMoe
[ -0.00035851384745910764, 0.041463855654001236, 0.00232943007722497, -0.0018635441083461046, -0.0012811865890398622, 0.054275721311569214, 0.05334394797682762, -0.01584012433886528, -0.005095628555864096, -0.012753630056977272, 0.014093051664531231, -0.030515534803271294, -0.00012921057350467...
[ "ACT2FN", "ModelNaiveMoe", "Module", "None", "Parameter", "__init__", "act_fn", "chunk", "class", "config", "continue", "current_hidden_states", "current_state", "def", "dim", "down_proj", "dtype", "empty", "expert_hit", "expert_idx", "expert_mask", "final_hidden_states", ...
glm4_moe/modeling_glm4_moe.py:Glm4MoeMoE
[ -0.0004175463109277189, 0.03760134056210518, -0.012415536679327488, -0.012297293171286583, -0.001625844044610858, 0.06952700763940811, 0.053445927798748016, -0.025777019560337067, -0.007922294549643993, -0.020810803398489952, 0.030743233859539032, -0.02885134145617485, -0.0007907514227554202...
[ "False", "ModelMLP", "ModelMoE", "ModelNaiveMoe", "ModelTopkRouter", "Module", "True", "__init__", "bool", "class", "config", "def", "denominator", "dim", "e_score_correction_bias", "expand", "experts", "forward", "gate", "gather", "group_idx", "group_mask", "group_scores...
glm4_moe/modeling_glm4_moe.py:Glm4MoeDecoderLayer
[ -0.0002400130615569651, 0.04778926819562912, 0.0105819096788764, -0.0029299373272806406, -0.0007822648040018976, 0.035500600934028625, 0.041872501373291016, -0.03686600551009178, 0.004835818894207478, -0.007139944471418858, -0.0037548711989074945, 0.015929756686091423, -0.0013085156679153442...
[ "False", "GradientCheckpointingLayer", "ModelAttention", "ModelDecoderLayer", "ModelMLP", "ModelMoE", "ModelRMSNorm", "None", "Tensor", "_", "__init__", "attention_mask", "cache_position", "class", "config", "def", "else", "eps", "first_k_dense_replace", "forward", "hidden_si...
glm4_moe/modeling_glm4_moe.py:Glm4MoePreTrainedModel
[ -0.00037101644556969404, 0.05074922740459442, 0.0014767908724024892, -0.005674369167536497, -0.0014840657822787762, 0.041670236736536026, 0.03398801386356354, -0.018972761929035187, -0.005121481604874134, -0.0031136281322687864, 0.008613401092588902, -0.009660976938903332, -0.002226098673418...
[ "ModelAttention", "ModelConfig", "ModelDecoderLayer", "ModelNaiveMoe", "ModelPreTrainedModel", "ModelTopkRouter", "PreTrainedModel", "True", "_can_compile_fullgraph", "_can_record_outputs", "_init_weights", "_keep_in_fp32_modules_strict", "_no_split_modules", "_skip_keys_device_placement",...
glm4_moe/modeling_glm4_moe.py:Glm4MoeModel
[ -0.00009217467595590279, 0.04854181781411171, 0.001228995737619698, -0.005983453243970871, -0.0004915982717648149, 0.03663109615445137, 0.028765521943569183, -0.014382760971784592, 0.009944330900907516, 0.01095561869442463, 0.009775782935321331, 0.0043541560880839825, -0.001334338216111064, ...
[ "BaseModelOutputWithPast", "DynamicCache", "Embedding", "False", "ModelDecoderLayer", "ModelModel", "ModelPreTrainedModel", "ModelRMSNorm", "ModelRotaryEmbedding", "ModuleList", "None", "ValueError", "You", "__init__", "_keys_to_ignore_on_load_unexpected", "and", "arange", "attenti...
glm4_moe/modeling_glm4_moe.py:Glm4MoeForCausalLM
[ -0.00028262686100788414, 0.03526614233851433, 0.008361488580703735, -0.001301150070503354, -0.0012087186332792044, 0.02741658128798008, 0.0389065183699131, -0.007394513580948114, 0.003640376031398773, 0.026734011247754097, 0.025823917239904404, 0.0026734010316431522, 0.0009385344455949962, ...
[ "CausalLMOutputWithPast", "GenerationMixin", "Linear", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "_pp_plan", "_tied_weights_keys", "_tp_plan", "attention_mask", "attentions", "auto_docstring", "cache_position", "can_return_tuple", "class", "colwi...
swin/modeling_swin.py:SwinEncoderOutput
[ -0.00021715721231885254, 0.02585153840482235, 0.03431620076298714, 0.020132170990109444, -0.0012082161847501993, 0.038434144109487534, 0.03729027137160301, -0.05284694954752922, 0.01509912870824337, 0.017615649849176407, -0.004146541003137827, 0.035917624831199646, -0.00303126429207623, 0....
[ "ModelEncoderOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "last_hidden_state", "r", "reshaped_hidden_states" ]
swin/modeling_swin.py:SwinModelOutput
[ -0.00020244609913788736, 0.025114018470048904, 0.02728295512497425, 0.01712319441139698, -0.0010916036553680897, 0.04520523175597191, 0.032648224383592606, -0.023059234023094177, 0.014954255893826485, 0.0061928885988891125, -0.00038705553743056953, 0.014954255893826485, -0.001319912844337523...
[ "ModelModelOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "last_hidden_state", "pooler_output", "r", "reshaped_hidden_states" ]
swin/modeling_swin.py:SwinMaskedImageModelingOutput
[ -0.00019999424694105983, 0.039871420711278915, 0.03443440794944763, -0.008155517280101776, -0.0010760752484202385, 0.05097198486328125, 0.011610285378992558, -0.04938618838787079, 0.015178324654698372, -0.00662635825574398, -0.012233276851475239, 0.03103627637028694, -0.0024070104118436575, ...
[ "ModelMaskedImageModelingOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "loss", "r", "reconstruction", "reshaped_hidden_states" ]
swin/modeling_swin.py:SwinImageClassifierOutput
[ -0.0002069334004772827, 0.034184690564870834, 0.02705344557762146, 0.017092345282435417, -0.0009692265302874148, 0.03327913582324982, 0.04437217861413956, -0.018903454765677452, 0.013470126315951347, -0.006197390146553516, 0.022638866677880287, 0.018450677394866943, -0.003339232876896858, ...
[ "ModelImageClassifierOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "logits", "loss", "r", "reshaped_hidden_states" ]
swin/modeling_swin.py:window_partition
[ -0.00004125074337935075, 0.009773793630301952, -0.008818882517516613, 0.00499923899769783, 0, 0.025501737371087074, 0.019772272557020187, -0.03999391198158264, 0.012245327234268188, 0.007920143194496632, -0.00150258035864681, -0.04403823986649513, 0.002359191421419382, -0.00117959571070969...
[ "Model_partition", "Model_size", "Models", "batch_size", "contiguous", "def", "height", "input_feature", "num_channels", "permute", "return", "shape", "view", "width" ]
swin/modeling_swin.py:window_reverse
[ -0.0000877177226357162, 0.022455736994743347, 0.02098877914249897, 0.02098877914249897, -0.0002803440729621798, 0.03836658596992493, 0.01602369174361229, -0.043783046305179596, 0.017152121290564537, 0.027646509930491447, -0.017942020669579506, -0.0253896526992321, -0.0025530706625431776, 0...
[ "Model_reverse", "Model_size", "Models", "contiguous", "def", "height", "num_channels", "permute", "return", "shape", "view", "width" ]
swin/modeling_swin.py:SwinEmbeddings
[ -0.00022775810793973505, 0.02710321545600891, 0.011615663766860962, 0.005551604088395834, -0.0011316731106489897, 0.017992891371250153, 0.023231327533721924, -0.026761578395962715, 0.0014448404544964433, 0.005864771082997322, 0.02175089903175831, 0.030747344717383385, -0.002405694918707013, ...
[ "Dropout", "False", "LayerNorm", "ModelEmbeddings", "ModelPatchEmbeddings", "Module", "None", "Parameter", "_", "__init__", "align_corners", "and", "batch_size", "bicubic", "bool_masked_pos", "cat", "class", "class_pos_embed", "config", "def", "dim", "dropout", "else", ...
swin/modeling_swin.py:SwinPatchEmbeddings
[ -0.00009367455641040578, 0.0035018527414649725, 0.020058611407876015, 0.009356950409710407, -0.0003029102517757565, -0.0034598305355757475, 0.014791825786232948, -0.020506849512457848, 0.004006119444966316, 0.0035018527414649725, 0.007732090540230274, 0.010757691226899624, -0.002395267132669...
[ "Conv2d", "Iterable", "ModelPatchEmbeddings", "Module", "_", "__init__", "abc", "class", "collections", "config", "def", "else", "embed_dim", "embeddings", "flatten", "forward", "functional", "grid_size", "height", "hidden_size", "if", "image_size", "isinstance", "kerne...
swin/modeling_swin.py:SwinPatchMerging
[ -0.00021154274872969836, 0.02943691983819008, 0.017414972186088562, 0.015842005610466003, -0.00020013173343613744, 0.013257849030196667, -0.0029633541125804186, -0.03865000233054161, 0.0009409702615812421, 0.011347820051014423, 0.005112136714160442, -0.0016291425563395023, -0.000811060192063...
[ "LayerNorm", "Linear", "ModelPatchMerging", "Module", "__init__", "batch_size", "cat", "class", "def", "dim", "forward", "functional", "height", "if", "input_dimensions", "input_feature", "input_feature_0", "input_feature_1", "input_feature_2", "input_feature_3", "input_resol...
swin/modeling_swin.py:drop_path
[ 0, 0.017753854393959045, 0.03775503486394882, -0.02123720571398735, 0, 0.043598074465990067, 0.0339345820248127, -0.027304979041218758, 0.014832334592938423, 0.0017486985307186842, 0.020113544538617134, -0.04944111406803131, 0.0016574009787291288, -0.018315685912966728, -0.05034004524350...
[ "False", "Model_path", "Model_prob", "def", "device", "div", "dtype", "floor_", "if", "input", "keep_prob", "ndim", "not", "or", "output", "rand", "random_tensor", "return", "shape", "torch", "training" ]
swin/modeling_swin.py:SwinDropPath
[ -0.00011107319733127952, 0.018094833940267563, 0.03596488758921623, -0.025063030421733856, -0.00047414645086973906, 0.031918834894895554, 0.033717080950737, -0.026861274614930153, 0.0066872211173176765, -0.002430439693853259, 0.009721758775413036, -0.033717080950737, -0.000818341679405421, ...
[ "ModelDropPath", "Module", "None", "__init__", "class", "def", "drop_path", "drop_prob", "extra_repr", "f", "forward", "hidden_states", "nn", "p", "return", "self", "super", "training" ]
swin/modeling_swin.py:SwinSelfAttention
[ 0, 0.034425679594278336, 0.04403804615139961, -0.011903684586286545, -0.00010434934665681794, 0.012127228081226349, 0.022019023075699806, -0.0135802598670125, 0.006370986346155405, 0.02604280225932598, 0.017101068049669266, 0.0326373316347599, 0.001201545586809516, -0.00419143820181489, ...
[ "Dropout", "False", "Iterable", "Linear", "ModelSelfAttention", "Module", "None", "Parameter", "The", "ValueError", "__init__", "a", "abc", "all_head_size", "arange", "attention", "attention_head_size", "attention_mask", "attention_probs", "attention_probs_dropout_prob", "att...
swin/modeling_swin.py:SwinSelfOutput
[ -0.00004677825199905783, 0.04611024260520935, 0.04879627376794815, 0.0014549348270520568, -0.00040045680361799896, 0.04051433876156807, 0.03379925340414047, -0.024510055780410767, 0.0021124533377587795, 0.014885102398693562, 0.010744133964180946, 0.011583519168198109, 0.002350279362872243, ...
[ "Dropout", "Linear", "ModelSelfOutput", "Module", "__init__", "attention_probs_dropout_prob", "class", "config", "def", "dense", "dim", "dropout", "forward", "hidden_states", "input_tensor", "nn", "return", "self", "super" ]
swin/modeling_swin.py:SwinAttention
[ 0.00015754895866848528, 0.031094355508685112, 0.04551492631435394, 0.0031685824505984783, 0.0005386589909903705, 0.03650207072496414, 0.04461364075541496, -0.018589017912745476, 0.014758553355932236, 0.0018589016981422901, 0.015096534974873066, 0.02320810593664646, 0.0020138102117925882, -...
[ "False", "ModelAttention", "ModelSelfAttention", "ModelSelfOutput", "Module", "None", "__init__", "attention_mask", "attention_output", "class", "config", "def", "dim", "forward", "hidden_states", "nn", "num_heads", "output", "output_attentions", "outputs", "return", "self"...
swin/modeling_swin.py:SwinIntermediate
[ -0.0002936642267741263, 0.023951541632413864, 0.039881035685539246, 0.0161586944013834, -0.0009311304893344641, 0.04263145104050636, 0.028764769434928894, -0.018565310165286064, -0.004383475985378027, -0.002549865050241351, 0.02097192406654358, -0.025326749309897423, -0.0019768616184592247, ...
[ "ACT2FN", "Linear", "ModelIntermediate", "Module", "__init__", "class", "config", "def", "dense", "dim", "else", "forward", "hidden_act", "hidden_states", "if", "int", "intermediate_act_fn", "isinstance", "mlp_ratio", "nn", "return", "self", "str", "super" ]
swin/modeling_swin.py:SwinOutput
[ -0.00027769507141783834, 0.044259220361709595, 0.04999227821826935, 0.024652156978845596, -0.0011824435787275434, 0.03600361570715904, 0.04609379917383194, -0.021097660064697266, -0.0009244558750651777, 0.00905823428183794, 0.007968952879309654, -0.007309651002287865, -0.00014690964599139988...
[ "Dropout", "Linear", "ModelOutput", "Module", "__init__", "class", "config", "def", "dense", "dim", "dropout", "forward", "hidden_dropout_prob", "hidden_states", "int", "mlp_ratio", "nn", "return", "self", "super" ]
swin/modeling_swin.py:SwinLayer
[ -0.0001743055909173563, 0.02107160910964012, 0.021296974271535873, -0.010761169716715813, -0.000531720113940537, 0.04304467886686325, 0.022085750475525856, -0.035833004862070084, 0.002000112785026431, 0.01994478516280651, 0.01498675998300314, -0.013352865353226662, 0.0005422840476967394, 0...
[ "False", "Identity", "LayerNorm", "ModelAttention", "ModelDropPath", "ModelIntermediate", "ModelLayer", "ModelOutput", "Module", "None", "_", "__init__", "always_partition", "attention", "attention_output", "attention_outputs", "attention_windows", "attn_mask", "batch_size", "c...
swin/modeling_swin.py:SwinStage
[ -0.00017466909775976092, -0.002935852389782667, 0.0011080019176006317, 0.009710895828902721, -0.00009924380719894543, 0.037714410573244095, 0.0231480672955513, -0.027326010167598724, 0.001164460671134293, 0.014171133749186993, 0.022809313610196114, -0.0287939365953207, -0.001277378061786294,...
[ "False", "GradientCheckpointingLayer", "LayerNorm", "ModelLayer", "ModelStage", "ModuleList", "None", "__init__", "always_partition", "blocks", "class", "config", "def", "depth", "dim", "downsample", "drop_path", "drop_path_rate", "else", "enumerate", "for", "forward", "h...
swin/modeling_swin.py:SwinEncoder
[ -0.00025126183754764497, 0.022125298157334328, 0.018475765362381935, 0.024634353816509247, -0.0008232835098169744, 0.02634507231414318, 0.011975033208727837, -0.04721584543585777, 0.005046620965003967, 0.026002928614616394, 0.003592509776353836, -0.015966709703207016, -0.0022667027078568935,...
[ "False", "ModelEncoder", "ModelEncoderOutput", "ModelPatchMerging", "ModelStage", "Module", "ModuleList", "None", "True", "_", "__init__", "all_hidden_states", "all_reshaped_hidden_states", "all_self_attentions", "always_partition", "and", "attentions", "batch_size", "class", "...
swin/modeling_swin.py:SwinPreTrainedModel
[ -0.00024896516697481275, 0.03414379432797432, -0.006800305563956499, 0.012121045961976051, -0.0012163725914433599, 0.01912052370607853, 0.016047582030296326, -0.0023331590928137302, -0.0007148856529965997, 0.019461961463093758, 0.004524052608758211, 0.007739259861409664, -0.00217666686512529...
[ "Model", "ModelConfig", "ModelEmbeddings", "ModelPreTrainedModel", "ModelSelfAttention", "ModelStage", "None", "PreTrainedModel", "True", "_init_weights", "_no_split_modules", "base_model_prefix", "class", "config", "copy_", "create_relative_position_index", "def", "elif", "if", ...
swin/modeling_swin.py:SwinModel
[ -0.000011052514309994876, 0.036978866904973984, 0.0013026646338403225, 0.026781665161252022, 0.00009454824612475932, 0.036530639976263046, 0.014679490588605404, -0.01378303300589323, 0.008684431202709675, 0.016472404822707176, 0.024988749995827675, -0.001099561108276248, -0.00018734559125732...
[ "AdaptiveAvgPool1d", "False", "LayerNorm", "ModelEmbeddings", "ModelEncoder", "ModelModel", "ModelModelOutput", "ModelPreTrainedModel", "None", "True", "__init__", "add_pooling_layer", "attentions", "auto_docstring", "bool_masked_pos", "class", "config", "def", "depths", "else"...
swin/modeling_swin.py:SwinForMaskedImageModeling
[ -0.0001356861466774717, 0.05692164972424507, 0.0139502864331007, -0.0011275081196799874, -0.0006793061038479209, 0.03383924812078476, 0.0061347647570073605, -0.060059066861867905, 0.010756847448647022, 0.012101453728973866, 0.013334008865058422, 0.019720887765288353, -0.001281577511690557, ...
[ "Conv2d", "False", "Model", "ModelForMaskedImageModeling", "ModelMaskedImageModelingOutput", "ModelModel", "ModelPreTrainedModel", "None", "PixelShuffle", "Sequential", "True", "__init__", "add_pooling_layer", "attentions", "auto_docstring", "batch_size", "bool_masked_pos", "class"...
swin/modeling_swin.py:SwinForImageClassification
[ -0.00009624342783354223, 0.038351044058799744, 0.012876586057245731, 0.022631576284766197, -0.0001829060638556257, 0.00443155225366354, 0.026422087103128433, -0.006298936437815428, 0.003219146514311433, 0.010925588198006153, 0.04370235279202461, 0.00468239514157176, 0.00018116409773938358, ...
[ "False", "Identity", "Linear", "Model", "ModelForImageClassification", "ModelImageClassifierOutput", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "attentions", "auto_docstring", "class", "classifier", "config", "def", "else", "forward", "hidden_states", "if", "in...
swin/modeling_swin.py:SwinBackbone
[ -0.00023383869847748429, 0.003968171775341034, 0.028003955259919167, 0.04126898571848869, -0.0008963816799223423, 0.015759311616420746, 0.005385376047343016, -0.008389849215745926, -0.0020974623039364815, 0.0027352040633559227, 0.004875182639807463, 0.004648429807275534, -0.00240924721583724...
[ "BackboneMixin", "BackboneOutput", "LayerNorm", "ModelBackbone", "ModelEmbeddings", "ModelEncoder", "ModelPreTrainedModel", "ModuleDict", "None", "True", "__init__", "_init_backbone", "_out_features", "always_partition", "attentions", "batch_size", "channels", "class", "config", ...
jamba/modeling_jamba.py:JambaRMSNorm
[ -0.00009884802420856431, 0.04179859161376953, 0.03230918198823929, 0.053095508366823196, -0.00046952810953371227, 0.03931327164173126, 0.022028988227248192, -0.02937198430299759, 0.007964326068758965, 0.042476408183574677, 0.019995542243123055, 0.006975846365094185, 0.0028524715453386307, ...
[ "ModelRMSNorm", "Module", "Parameter", "True", "__init__", "class", "def", "eps", "extra_repr", "f", "float32", "forward", "hidden_size", "hidden_states", "keepdim", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "shape", "super", "to", "torch", "tuple", ...
jamba/modeling_jamba.py:HybridMambaAttentionDynamicCache
[ -0.00007408379315165803, -0.017843235284090042, 0.03187093511223793, -0.032319821417331696, 0, 0.024464309215545654, -0.022219877690076828, -0.05745745822787285, 0.0004979833611287177, 0.029626501724123955, 0.0026933185290545225, 0.01077327411621809, 0.002454847563058138, 0.001409783842973...
[ "Any", "False", "ModelMambaAttentionDynamicCache", "None", "_", "__getitem__", "__init__", "__len__", "append", "batch_size", "beam_idx", "cache_kwargs", "cache_position", "cat", "class", "config", "conv_kernel_size", "conv_states", "def", "device", "dim", "dtype", "else"...
jamba/modeling_jamba.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
jamba/modeling_jamba.py:apply_rotary_pos_emb
[ -0.0001444592053303495, 0.027112245559692383, 0.028019767254590988, 0.0028360087890177965, -0.0005707467789761722, 0.021667107939720154, 0.046510547399520874, -0.002183726755902171, 0.01293220091611147, 0.03652779385447502, 0.007884104736149311, 0.0017654155381023884, -0.0007834474672563374,...
[ "Model_rotary_pos_emb", "cos", "def", "k", "k_embed", "q", "q_embed", "return", "rotate_half", "sin", "unsqueeze", "unsqueeze_dim" ]
jamba/modeling_jamba.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
jamba/modeling_jamba.py:eager_attention_forward
[ 0, 0.021594731137156487, 0.01775064319372177, -0.018768195062875748, -0.00008832923776935786, 0.039797618985176086, 0.05359111353754997, -0.035049039870500565, 0.02069023996591568, 0.011645326390862465, 0.029395969584584236, 0.022499222308397293, 0.002741739386692643, -0.014584923163056374...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "causal_mask", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "key_states", "kwargs", "matmul", "module", "nn", "not", "num_key_value_groups", ...
jamba/modeling_jamba.py:JambaAttention
[ -0.00010017884778790176, 0.03014504536986351, 0.03261964023113251, -0.0035009870771318674, -0.00028471884434111416, 0.0287952683866024, 0.04206808656454086, -0.024858415126800537, 0.004021214321255684, 0.016197338700294495, 0.02350863628089428, 0.016984708607196808, -0.0005799828213639557, ...
[ "ALL_ATTENTION_FUNCTIONS", "Linear", "ModelAttention", "Module", "None", "True", "__init__", "_attn_implementation", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "cache_position", "class", "config", "contiguous", "def", "dropout", ...
jamba/modeling_jamba.py:JambaMambaMixer
[ -0.00019614165648818016, 0.015361742116510868, 0.015705661848187447, -0.00227846740745008, -0.0006699267541989684, 0.05571497604250908, 0.00495817419141531, -0.03943611681461334, 0.012381105683743954, 0.007967471145093441, 0.023959733545780182, -0.012667705304920673, 0.003754455829039216, ...
[ "A", "ACT2FN", "A_log", "B", "C", "CUDA", "Conv1d", "D", "False", "Fast", "HybridMambaAttentionDynamicCache", "Linear", "Make", "Mamba", "ModelMambaMixer", "ModelRMSNorm", "Module", "None", "Parameter", "The", "To", "True", "ValueError", "_", "__init__", "a", "act...
jamba/modeling_jamba.py:JambaMLP
[ -0.0002434849739074707, 0.024471141397953033, 0.0228551235049963, 0.026548879221081734, -0.0009450823999941349, 0.060485273599624634, 0.036244992166757584, -0.005078915972262621, 0, -0.005021201446652412, 0.028164898976683617, -0.04640282690525055, -0.0015583038330078125, 0.009176678024232...
[ "ACT2FN", "Linear", "ModelMLP", "Module", "__init__", "act_fn", "class", "config", "def", "down_proj", "forward", "gate_proj", "hidden_act", "hidden_size", "intermediate_size", "nn", "return", "self", "super", "up_proj", "x" ]
jamba/modeling_jamba.py:JambaExperts
[ -0.00033012471976689994, 0.03738027438521385, -0.010854209773242474, 0.006558961234986782, -0.001262454898096621, 0.05874042958021164, 0.0631517693400383, -0.01555576641112566, -0.005252973176538944, -0.018806224688887596, 0.011492692865431309, -0.016020117327570915, -0.0018283829558640718, ...
[ "ACT2FN", "ModelExperts", "Module", "None", "Parameter", "__init__", "act_fn", "chunk", "class", "config", "continue", "current_hidden_states", "current_state", "def", "dim", "down_proj", "dtype", "empty", "expert_hit", "expert_idx", "expert_mask", "final_hidden_states", ...
jamba/modeling_jamba.py:JambaSparseMoeBlock
[ -0.0003686461132019758, 0.03363804519176483, 0.00224837614223361, -0.01833740621805191, -0.0012920863227918744, 0.052559442818164825, 0.05115785822272301, -0.03574042022228241, -0.008759907446801662, -0.020089387893676758, 0.02300935611128807, -0.017286216840147972, -0.0008212412940338254, ...
[ "Linear", "ModelExperts", "ModelSparseMoeBlock", "Module", "__init__", "batch_size", "class", "config", "def", "dim", "dtype", "experts", "ffn_dim", "float", "forward", "functional", "hidden_dim", "hidden_size", "hidden_states", "intermediate_size", "nn", "num_experts", "...
jamba/modeling_jamba.py:JambaAttentionDecoderLayer
[ -0.00020280387252569199, 0.04808038845658302, 0.02562030218541622, -0.0015871607465669513, -0.0006736615905538201, 0.040857043117284775, 0.04582309350371361, -0.04356579855084419, 0.005671454593539238, -0.008013398386538029, 0.010383558459579945, 0.02370160073041916, 0.000031743216823088005,...
[ "False", "GradientCheckpointingLayer", "ModelAttention", "ModelAttentionDecoderLayer", "ModelMLP", "ModelRMSNorm", "ModelSparseMoeBlock", "None", "_", "__init__", "attention_mask", "cache_position", "class", "config", "def", "else", "eps", "feed_forward", "ffn_layer_class", "fo...
jamba/modeling_jamba.py:JambaMambaDecoderLayer
[ -0.00018875245586968958, 0.022458884865045547, 0.016106875613331795, 0.010945870541036129, -0.0005033398629166186, 0.06215893104672432, 0.02393345721065998, -0.037885189056396484, 0.0056997924111783504, -0.010775727219879627, 0.018262021243572235, -0.0028073606081306934, 0.000900340382941067...
[ "GradientCheckpointingLayer", "ModelMLP", "ModelMambaDecoderLayer", "ModelMambaMixer", "ModelRMSNorm", "ModelSparseMoeBlock", "None", "__init__", "attention_mask", "cache_params", "class", "config", "def", "else", "eps", "feed_forward", "ffn_layer_class", "forward", "hidden_size"...
jamba/modeling_jamba.py:JambaPreTrainedModel
[ -0.00032878422643989325, 0.027481330558657646, -0.01000366359949112, 0.008106417022645473, -0.0012576445005834103, 0.055422596633434296, 0.017362680286169052, -0.016212834045290947, -0.0030614661518484354, -0.012245864607393742, 0.010061156004667282, -0.019547389820218086, -0.003018346847966...
[ "A", "A_log", "D", "Linear", "ModelAttention", "ModelAttentionDecoderLayer", "ModelConfig", "ModelExperts", "ModelMambaDecoderLayer", "ModelMambaMixer", "ModelPreTrainedModel", "None", "OutputRecorder", "PreTrainedModel", "True", "_can_record_outputs", "_init_weights", "_is_statefu...
jamba/modeling_jamba.py:JambaModel
[ -0.00013742870942223817, 0.024356596171855927, -0.006455625407397747, -0.014208014123141766, -0.0004898100160062313, 0.05773415416479111, 0.010825153440237045, -0.02695012278854847, 0.008739056065678596, -0.013756966218352318, 0.025033168494701385, -0.019056780263781548, 0.001184001215733587...
[ "ALL_DECODER_LAYER_TYPES", "Embedding", "False", "HybridMambaAttentionDynamicCache", "ModelMambaDecoderLayer", "ModelModel", "ModelPreTrainedModel", "ModelRMSNorm", "ModuleList", "MoeModelOutputWithPast", "None", "True", "ValueError", "You", "__init__", "_update_mamba_mask", "all", ...
jamba/modeling_jamba.py:load_balancing_loss_func
[ -0.00027690583374351263, 0.01840798556804657, 0.001643570140004158, -0.028812499716877937, -0.0009432663209736347, 0.05808233842253685, 0.038874007761478424, -0.011547867208719254, 0, -0.02252405695617199, 0.031556546688079834, -0.0061455233953893185, -0.0008396499906666577, 0.014349081553...
[ "Model_balancing_loss_func", "None", "_", "attention_mask", "batch_size", "cat", "compute_device", "concatenated_gate_logits", "def", "device", "dim", "else", "expand", "expert_attention_mask", "expert_mask", "float", "for", "functional", "gate_logits", "if", "in", "is", ...
jamba/modeling_jamba.py:JambaForCausalLM
[ -0.00044288873323239386, 0.0379493422806263, -0.00161050446331501, -0.01030722912400961, -0.001698350184597075, 0.04966210201382637, 0.04029189422726631, -0.004597258288413286, -0.005153614562004805, 0.006207762751728296, 0.02857913449406624, -0.0060027893632650375, -0.00006588427640963346, ...
[ "GenerationMixin", "Linear", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "MoeCausalLMOutputWithPast", "None", "__init__", "_pp_plan", "_tied_weights_keys", "_tp_plan", "attention_mask", "attentions", "auto_docstring", "aux_loss", "cache_position", "can_return_tuple", ...
jamba/modeling_jamba.py:JambaForSequenceClassification
[ -0.0002121782599715516, 0.015340753830969334, -0.017272552475333214, 0.0278406273573637, -0.0007244244916364551, 0.027954263612627983, 0.015227118507027626, -0.011079433374106884, 0.0032101948745548725, -0.013863496482372284, 0.02909061498939991, 0.008409005589783192, -0.003437465289607644, ...
[ "GenericForSequenceClassification", "ModelForSequenceClassification", "ModelPreTrainedModel", "class", "pass" ]
m2m_100/modeling_m2m_100.py:shift_tokens_right
[ -0.0001357423752779141, 0.026739485561847687, 0.00459761219099164, -0.04174518957734108, -0.0005782272783108056, 0.05189942196011543, 0.04039129242300987, -0.057089366018772125, 0.006572046782821417, 0.007784913759678602, 0.020759768784046173, -0.009872172959148884, -0.0006698974757455289, ...
[ "Model_tokens_right", "Modeled_input_ids", "None", "clone", "decoder_start_token_id", "def", "if", "input_ids", "is", "masked_fill_", "new_zeros", "pad_token_id", "return", "shape" ]
m2m_100/modeling_m2m_100.py:M2M100ScaledWordEmbedding
[ -0.00018138372979592532, 0.03397626802325249, 0.021971320733428, 0.020159251987934113, -0.0008741811034269631, 0.019253218546509743, 0.04462216794490814, -0.020725524052977562, 0.00569102494046092, -0.00651211803779006, 0.005974160507321358, 0.018913455307483673, 0.002505749696865678, -0.0...
[ "Embedding", "ModelScaledWordEmbedding", "__init__", "class", "def", "embed_scale", "embedding_dim", "forward", "input_ids", "nn", "num_embeddings", "padding_idx", "return", "self", "super" ]