identifier
stringlengths
24
102
embedding
listlengths
2.56k
2.56k
tokens
listlengths
4
448
sam2_video/modeling_sam2_video.py:Sam2VideoMaskDownSampler
[ 0, 0.01759873516857624, 0.01748592220246792, -0.0008531438070349395, 0.00029436987824738026, 0.058211199939250946, 0.015906548127532005, -0.041966214776039124, 0.016696235164999962, 0.007896868512034416, 0.008122493512928486, -0.012691395357251167, 0.002044724766165018, 0.00065219670068472...
[ "ACT2FN", "Conv2d", "ModelMaskDownModelpler", "ModelMaskDownModelplerLayer", "Module", "ModuleList", "_", "__init__", "activation", "append", "class", "config", "def", "final_conv", "for", "forward", "in", "int", "kernel_size", "layer", "layers", "log2", "mask_downModelpl...
sam2_video/modeling_sam2_video.py:Sam2VideoMemoryEncoder
[ -0.00022780879226047546, 0.02249126136302948, 0.0025429818779230118, 0.0158229973167181, -0.0006145539227873087, 0.049729421734809875, 0.006187922321259975, -0.051763806492090225, 0.010341458953917027, 0.04272209480404854, -0.007402902469038963, 0.021813131868839264, 0.0015893636737018824, ...
[ "Conv2d", "ModelMaskDownModelpler", "ModelMemoryEncoder", "ModelMemoryFuser", "ModelPositionEmbeddingSine", "Module", "True", "__init__", "class", "config", "def", "device", "dtype", "feature_projection", "forward", "hidden_size", "kernel_size", "mask_downModelpler", "masks", "...
sam2_video/modeling_sam2_video.py:Sam2VideoPositionalEmbedding
[ -0.00034020061139017344, 0.034743890166282654, 0.016677068546414375, 0.013260585255920887, -0.0016141432570293546, 0.007991095073521137, 0.03173275291919708, -0.00634076027199626, 0.008801786229014397, 0.022699343040585518, 0.007064591161906719, 0.009149224497377872, 0.0016503348015248775, ...
[ "ModelPositionalEmbedding", "Module", "None", "__init__", "cat", "class", "clone", "config", "coordinates", "cos", "def", "dim", "dtype", "float32", "forward", "hidden_size", "if", "input_coords", "input_shape", "is", "nn", "not", "np", "pi", "positional_embedding", ...
sam2_video/modeling_sam2_video.py:Sam2VideoVisionEncoderOutput
[ -0.00023599078122060746, 0.027209196239709854, 0.00795523077249527, 0.03896910324692726, -0.0011817553313449025, 0.040583208203315735, 0.06410302221775055, -0.0548795647919178, 0.013546951115131378, 0.01948455162346363, 0.016256341710686684, 0.012394019402563572, 0.00024499805294908583, -0...
[ "BaseModelOutputWithPooling", "ModelVisionEncoderOutput", "None", "class", "fpn_hidden_states", "fpn_position_encoding", "r" ]
sam2_video/modeling_sam2_video.py:Sam2VideoMaskEmbedding
[ -0.000023781873096595518, 0.005298953503370285, 0.03021530993282795, 0.0033259389456361532, 0, 0.05344051122665405, -0.009244983084499836, -0.04284260421991348, 0.013360127806663513, 0.006088159512728453, 0.030666284263134003, 0.012965524569153786, 0.0017898061778396368, -0.001359970774501...
[ "ACT2FN", "Conv2d", "ModelLayerNorm", "ModelMaskEmbedding", "Module", "__init__", "activation", "channels_first", "class", "config", "conv1", "conv2", "conv3", "data_format", "def", "dense_embeddings", "eps", "forward", "hidden_act", "hidden_size", "hidden_states", "kernel_...
sam2_video/modeling_sam2_video.py:Sam2VideoPromptEncoder
[ -0.00017483846750110388, 0.010511500760912895, -0.003207138041034341, 0.012545985169708729, -0.0007488031405955553, 0.032099638134241104, 0.01718008704483509, -0.02283143252134323, 0.004266764968633652, 0.0014199003344401717, -0.00022781982261221856, 0.023170512169599533, -0.0042102518491446...
[ "Embedding", "ModelMaskEmbedding", "ModelPositionalEmbedding", "ModelPromptEncoder", "Module", "None", "Tensor", "__init__", "_embed_boxes", "_embed_points", "also", "batch_size", "be", "box_embeddings", "boxes", "cat", "clamp", "class", "config", "constant", "coords", "cor...
sam2_video/modeling_sam2_video.py:Sam2VideoTwoWayTransformer
[ -0.0002629486261866987, 0.04207178205251694, 0.04229919612407684, -0.018420618027448654, -0.0008670198149047792, 0.04298144206404686, 0.02353745698928833, -0.024901947006583214, 0.009551431983709335, 0.006452901754528284, 0.012280412018299103, 0.020467353984713554, -0.0004797035944648087, ...
[ "LayerNorm", "ModelAttention", "ModelTwoWayAttentionBlock", "ModelTwoWayTransformer", "Module", "ModuleList", "None", "_", "__init__", "append", "attention_similarity", "attn_out", "class", "config", "def", "final_attn_token_to_image", "flatten", "for", "forward", "hidden_size"...
sam2_video/modeling_sam2_video.py:Sam2VideoMaskDecoder
[ -0.0003465987101662904, 0.04436463490128517, 0.020680390298366547, -0.02010272443294525, -0.001270861946977675, 0.04551996290683746, 0.014499379321932793, -0.06331203132867813, 0.0004982356331311166, -0.01617460697889328, 0.019525060430169106, 0.014730445109307766, -0.0051989806815981865, ...
[ "Conv2d", "ConvTranspose2d", "Embedding", "GELU", "ModelFeedForward", "ModelLayerNorm", "ModelMaskDecoder", "ModelTwoWayTransformer", "Model_tokens_out", "Module", "ModuleList", "None", "True", "_", "__init__", "_dynamic_multimask_via_stability", "_get_stability_scores", "activatio...
sam2_video/modeling_sam2_video.py:get_1d_sine_pe
[ -0.00017746999219525605, 0.04678850620985031, 0.0166133102029562, 0.019325686618685722, -0.0008052369812503457, 0.011358079500496387, 0.05831610783934593, 0.009549827314913273, 0.014239979907870293, 0.017517436295747757, -0.005113961175084114, -0.014692042954266071, 0.0020907907746732235, ...
[ "Model_1d_sine_pe", "arange", "cat", "cos", "def", "device", "dim", "dim_t", "dtype", "float32", "pe_dim", "pos_embed", "pos_inds", "return", "sin", "temperature", "torch", "unsqueeze" ]
sam2_video/modeling_sam2_video.py:Sam2VideoModel
[ -0.0002554782258812338, 0.024180473759770393, -0.01704147644340992, 0.010247915051877499, -0.0012450065696612, 0.06540242582559586, 0.03868875652551651, -0.06171778216958046, 0.014853719621896744, 0.013874986208975315, 0.007945013232529163, 0.028786277398467064, -0.0019574668258428574, -0....
[ "AutoModel", "Cannot", "Conv2d", "Exactly", "F", "False", "Got", "Identity", "Linear", "ModelFeedForward", "ModelImageSegmentationOutput", "ModelMaskDecoder", "ModelMemoryAttention", "ModelMemoryEncoder", "ModelModel", "ModelPositionalEmbedding", "ModelPreTrainedModel", "ModelPromp...
canine/modeling_canine.py:CanineModelOutputWithPooling
[ -0.00020947544544469565, 0.0015397340757772326, 0.03047957271337509, 0.029792064800858498, -0.0010241022100672126, 0.03987553343176842, 0.04606311395764351, -0.007849062792956829, 0.014781447127461433, -0.011286608874797821, 0.00813552550971508, 0.00888032652437687, -0.0017545807640999556, ...
[ "ModelModelOutputWithPooling", "ModelOutput", "None", "attentions", "class", "hidden_states", "last_hidden_state", "pooler_output", "r" ]
canine/modeling_canine.py:CanineEmbeddings
[ -0.0002401918900432065, -0.013664249330759048, 0.01360731478780508, -0.020496373996138573, -0.0010817530564963818, 0.004497815389186144, 0.022546011954545975, -0.012468627654016018, 0.0048394217155873775, 0.003956939093768597, 0.033477410674095154, -0.0058073061518371105, -0.0003843070298898...
[ "Dropout", "Embedding", "Expected", "False", "HashBucketCodepointEmbedder_", "LayerNorm", "ModelEmbeddings", "Module", "None", "ValueError", "_PRIMES", "__init__", "_embed_hash_buckets", "_hash_bucket_tensors", "append", "arange", "be", "cat", "char_position_embeddings", "class...
canine/modeling_canine.py:CharactersToMolecules
[ -0.0004377936420496553, 0.02848772704601288, 0.02415008842945099, 0.006740923039615154, -0.0015093805268406868, 0.030246227979660034, 0.011606111191213131, -0.0295428279787302, -0.0002454575151205063, 0.012778445146977901, 0.03048069402575493, 0.024384556338191032, -0.0005751765565946698, ...
[ "ACT2FN", "Conv1d", "LayerNorm", "ModelToMolecules", "Module", "__init__", "activation", "cat", "char_encoding", "class", "cls_encoding", "config", "conv", "def", "dim", "downsampled", "downsampled_truncated", "downsampling_rate", "eps", "forward", "hidden_act", "hidden_siz...
canine/modeling_canine.py:ConvProjection
[ -0.0004426336963661015, 0.015163418836891651, 0.015398510731756687, 0.017749428749084473, -0.001638295711018145, 0.035498857498168945, 0.02503727190196514, -0.0345584899187088, -0.004819381050765514, 0.021158259361982346, 0.02162844128906727, 0.006905820686370134, -0.0023949972819536924, 0...
[ "ACT2FN", "ConstantPad1d", "Dropout", "LayerNorm", "Model", "Model1d", "ModelProjection", "Module", "None", "__init__", "activation", "class", "config", "def", "dropout", "else", "eps", "final_char_seq", "final_seq_char_positions", "forward", "hidden_act", "hidden_dropout_p...
canine/modeling_canine.py:CanineSelfAttention
[ -0.000033976375561906025, 0.04085231572389603, 0.04511711746454239, -0.014926807954907417, -0.0003857962437905371, 0.019528305158019066, 0.0323227122426033, -0.018518220633268356, 0.005218771286308765, 0.0207628533244133, 0.004405091982334852, 0.01762036792933941, -0.0006768970633856952, -...
[ "Dropout", "False", "Linear", "ModelSelfAttention", "Module", "None", "The", "ValueError", "_", "__init__", "a", "all_head_size", "and", "attention", "attention_head_size", "attention_mask", "attention_probs", "attention_probs_dropout_prob", "attention_scores", "batch_size", ...
canine/modeling_canine.py:CanineSelfOutput
[ -0.00012168083776487038, 0.04875698313117027, 0.03859927877783775, 0.020879726856946945, -0.0006630723946727812, 0.05620596557855606, 0.02370131015777588, -0.01952536590397358, 0.0035551965702325106, 0.017155233770608902, 0.01365646906197071, 0.003512872848659754, 0.0036962758749723434, -0...
[ "Dropout", "LayerNorm", "Linear", "ModelSelfOutput", "Module", "__init__", "class", "config", "def", "dense", "dropout", "eps", "forward", "hidden_dropout_prob", "hidden_size", "hidden_states", "input_tensor", "layer_norm_eps", "nn", "return", "self", "super" ]
canine/modeling_canine.py:CanineAttention
[ -0.0002279965701745823, 0.03100753389298916, 0.03260350972414017, -0.01823972538113594, -0.0011257331352680922, 0.0330595038831234, 0.019151711836457253, -0.03442748263478279, 0.0023939639795571566, 0.0017883481923490763, 0.02234366536140442, 0.03739143908023834, 0.003989940043538809, -0.0...
[ "Check", "Expected", "False", "ModelAttention", "ModelSelfAttention", "ModelSelfOutput", "Module", "None", "ValueError", "__init__", "always_attend_to_first_position", "and", "append", "attend_from_chunk_stride", "attend_from_chunk_width", "attend_to_chunk_stride", "attend_to_chunk_w...
canine/modeling_canine.py:CanineIntermediate
[ -0.00025367451598867774, 0.02240910567343235, 0.04047359153628349, 0.012862369418144226, -0.0009432404185645282, 0.03635763004422188, 0.03452831506729126, -0.018864808604121208, -0.001329111517407, -0.003772961674258113, 0.023209432139992714, -0.02103712037205696, -0.0013719861162826419, 0...
[ "ACT2FN", "Linear", "ModelIntermediate", "Module", "__init__", "class", "config", "def", "dense", "else", "forward", "hidden_act", "hidden_size", "hidden_states", "if", "intermediate_act_fn", "intermediate_size", "isinstance", "nn", "return", "self", "str", "super" ]
canine/modeling_canine.py:CanineOutput
[ -0.0002591986849438399, 0.03935529664158821, 0.0507957898080349, 0.030660521239042282, -0.0012513039400801063, 0.04988054931163788, 0.03798243775963783, -0.023681821301579475, 0.002259497530758381, 0.014758236706256866, 0.011040075682103634, 0.00267421524040401, 0.0013013561256229877, 0.00...
[ "Dropout", "LayerNorm", "Linear", "ModelOutput", "Module", "__init__", "class", "config", "def", "dense", "dropout", "eps", "forward", "hidden_dropout_prob", "hidden_size", "hidden_states", "input_tensor", "intermediate_size", "layer_norm_eps", "nn", "return", "self", "su...
canine/modeling_canine.py:CanineLayer
[ -0.0001648626202950254, 0.015234364196658134, 0.020425332710146904, -0.00009785962902242318, -0.000603026885073632, 0.04017358273267746, 0.025052065029740334, -0.009253465570509434, 0.0035546848084777594, -0.0019466131925582886, 0.007504334673285484, 0.013090267777442932, 0.00245442520827054...
[ "False", "GradientCheckpointingLayer", "ModelAttention", "ModelIntermediate", "ModelLayer", "ModelOutput", "None", "__init__", "always_attend_to_first_position", "apply_chunking_to_forward", "attend_from_chunk_stride", "attend_from_chunk_width", "attend_to_chunk_stride", "attend_to_chunk_w...
canine/modeling_canine.py:CanineEncoder
[ -0.00023000997316557914, 0.019513403996825218, 0.012096027843654156, 0.0102131562307477, -0.0009343040874227881, 0.04610184207558632, 0.015975886955857277, -0.04199375957250595, 0.0063618263229727745, 0.012495425529778004, 0.011696631088852882, 0.008900851011276245, 0.0006383222644217312, ...
[ "BaseModelOutput", "False", "ModelEncoder", "ModelLayer", "Module", "ModuleList", "None", "True", "_", "__init__", "all_hidden_states", "all_self_attentions", "always_attend_to_first_position", "attend_from_chunk_stride", "attend_from_chunk_width", "attend_to_chunk_stride", "attend_t...
canine/modeling_canine.py:CaninePooler
[ -0.0002915378427132964, 0.011618588119745255, 0.03685896843671799, 0.022893769666552544, -0.001151842763647437, 0.03296703100204468, 0.037087906152009964, -0.0066105760633945465, -0.0011303798528388143, 0.0015954095870256424, 0.014022434130311012, -0.016025640070438385, -0.001244848710484802...
[ "Linear", "ModelPooler", "Module", "Tanh", "__init__", "activation", "class", "config", "def", "dense", "first_token_tensor", "forward", "hidden_size", "hidden_states", "nn", "pooled_output", "return", "self", "super" ]
canine/modeling_canine.py:CaninePredictionHeadTransform
[ -0.0002919524849858135, 0.041955187916755676, 0.0442478209733963, 0.03209686279296875, -0.0013899088371545076, 0.026136018335819244, 0.029574967920780182, -0.011348534375429153, -0.0021780014503747225, 0.020519066601991653, 0.01662158966064453, 0.00664863595739007, 0.0001226916938321665, 0...
[ "ACT2FN", "LayerNorm", "Linear", "ModelPredictionHeadTransform", "Module", "__init__", "class", "config", "def", "dense", "else", "eps", "forward", "hidden_act", "hidden_size", "hidden_states", "if", "isinstance", "layer_norm_eps", "nn", "return", "self", "str", "super"...
canine/modeling_canine.py:CanineLMPredictionHead
[ -0.0003514417330734432, 0.03737177327275276, 0.032988663762807846, 0.019608646631240845, -0.0016941294306889176, 0.021684857085347176, 0.05328938364982605, -0.028028830885887146, -0.0002207775105489418, 0.0032008232083171606, 0.020762097090482712, 0.005680740345269442, 0.0002793511375784874,...
[ "Linear", "ModelLMPredictionHead", "ModelPredictionHeadTransform", "Module", "Parameter", "__init__", "bias", "class", "config", "decoder", "def", "forward", "hidden_size", "hidden_states", "nn", "return", "self", "super", "torch", "transform", "vocab_size", "zeros" ]
canine/modeling_canine.py:CanineOnlyMLMHead
[ -0.000383650854928419, 0.01940907910466194, 0.01812293566763401, 0.030399763956665993, -0.0018342165276408195, 0.03998738154768944, 0.041156601160764694, -0.00865224003791809, -0.0024407501332461834, -0.00499842269346118, 0.015082959085702896, 0.010698378086090088, -0.0020753685384988785, ...
[ "ModelLMPredictionHead", "ModelOnlyMLMHead", "Module", "__init__", "class", "config", "def", "forward", "nn", "prediction_scores", "predictions", "return", "self", "sequence_output", "super" ]
canine/modeling_canine.py:CaninePreTrainedModel
[ -0.0003105697687715292, 0.03148064389824867, -0.010340358130633831, 0.0037052948027849197, -0.0013715336099267006, 0.02906789444386959, 0.024472180753946304, 0.009019089862704277, -0.003906357567757368, -0.00505528599023819, 0.00404997356235981, -0.0027287055272608995, -0.001709031406790018,...
[ "Model", "ModelConfig", "ModelEmbeddings", "ModelPreTrainedModel", "PreTrainedModel", "True", "_init_weights", "arange", "base_model_prefix", "class", "config", "copy_", "def", "expand", "if", "init", "isinstance", "module", "position_ids", "self", "shape", "super", "supp...
canine/modeling_canine.py:CanineModel
[ -0.00030296845943666995, 0.026084139943122864, 0.012753528542816639, -0.003303798846900463, -0.0011397384805604815, 0.05193744972348213, 0.0027988513465970755, -0.011137696914374828, 0.0012190872803330421, -0.0073289512656629086, 0.04270412400364876, 0.02250622771680355, 0.002553591271862387...
[ "CharactersToMolecules", "ConvProjection", "False", "MaxPool1d", "ModelEmbeddings", "ModelEncoder", "ModelModel", "ModelModelOutputWithPooling", "ModelPooler", "ModelPreTrainedModel", "None", "True", "__init__", "_create_3d_attention_mask_from_input_mask", "_downsample_attention_mask", ...
canine/modeling_canine.py:CanineForSequenceClassification
[ -0.0003463077009655535, 0.02970391884446144, 0.002870426746085286, 0.011767322197556496, -0.001156739192083478, 0.02719051018357277, 0.020107269287109375, 0.012052936479449272, -0.0060835909098386765, 0.010053634643554688, 0.05895085632801056, 0.000642632890958339, -0.00012227875413373113, ...
[ "BCEWithLogitsLoss", "CrossEntropyLoss", "Dropout", "Linear", "MSELoss", "Model", "ModelForSequenceClassification", "ModelModel", "ModelPreTrainedModel", "None", "SequenceClassifierOutput", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "class", "classifier", ...
canine/modeling_canine.py:CanineForMultipleChoice
[ -0.00023257160501088947, 0.05476617440581322, 0.017838774248957634, 0.03135988488793373, -0.0008983300067484379, 0.04067695140838623, 0.04476737231016159, -0.0015268060378730297, -0.00023434696777258068, 0.010226049460470676, 0.04181317985057831, -0.007612725719809532, -0.001185937668196857,...
[ "CrossEntropyLoss", "Dropout", "Linear", "Model", "ModelForMultipleChoice", "ModelModel", "ModelPreTrainedModel", "MultipleChoiceModelOutput", "None", "__init__", "attention_mask", "attentions", "auto_docstring", "class", "classifier", "config", "def", "dropout", "else", "forwa...
canine/modeling_canine.py:CanineForTokenClassification
[ -0.0002629574737511575, 0.036842476576566696, 0.000557896273676306, -0.03138433024287224, -0.0008990303031168878, 0.0375247448682785, 0.040481239557266235, 0.01580587774515152, -0.00210365979000926, 0.009438041597604752, 0.05731051787734032, 0.020695464685559273, -0.0008279607282020152, -0...
[ "CrossEntropyLoss", "Dropout", "Linear", "Model", "ModelForTokenClassification", "ModelModel", "ModelPreTrainedModel", "None", "TokenClassifierOutput", "__init__", "attention_mask", "attentions", "auto_docstring", "class", "classifier", "config", "def", "dropout", "else", "forw...
canine/modeling_canine.py:CanineForQuestionAnswering
[ -0.00018441778956912458, 0.027989352121949196, 0.01326403021812439, 0.01292680948972702, -0.0006041878368705511, 0.04766058176755905, 0.037768762558698654, 0.024841954931616783, 0.0026134636718779802, 0.03147397190332413, 0.023043442517518997, 0.02169455774128437, 0.0007657729438506067, 0....
[ "CrossEntropyLoss", "Linear", "Model", "ModelForQuestionAnswering", "ModelModel", "ModelPreTrainedModel", "None", "QuestionAnsweringModelOutput", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "clamp_", "class", "config", "def", "dim", "else", "end_logits"...
deepseek_v3/modeling_deepseek_v3.py:DeepseekV3RMSNorm
[ -0.00009884802420856431, 0.04179859161376953, 0.03230918198823929, 0.053095508366823196, -0.00046952810953371227, 0.03931327164173126, 0.022028988227248192, -0.02937198430299759, 0.007964326068758965, 0.042476408183574677, 0.019995542243123055, 0.006975846365094185, 0.0028524715453386307, ...
[ "ModelRMSNorm", "Module", "Parameter", "True", "__init__", "class", "def", "eps", "extra_repr", "f", "float32", "forward", "hidden_size", "hidden_states", "keepdim", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "shape", "super", "to", "torch", "tuple", ...
deepseek_v3/modeling_deepseek_v3.py:DeepseekV3RotaryEmbedding
[ -0.00029968636226840317, 0.04852752387523651, 0.0020364229567348957, -0.005228263325989246, -0.0013792794197797775, 0.04136393964290619, 0.040670689195394516, 0.0062681385315954685, -0.0026430170983076096, 0.021144136786460876, 0.0021808501332998276, -0.004043960478156805, -0.001364836702123...
[ "False", "ModelRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "arange", "attention_factor", "attention_scaling", "base", "cat", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cpu", "def", "default", "device", ...
deepseek_v3/modeling_deepseek_v3.py:DeepseekV3MLP
[ -0.00024263472005259246, 0.02898676134645939, 0.02358049899339676, 0.02829659916460514, -0.0008483228739351034, 0.05774346739053726, 0.029906975105404854, -0.005061180330812931, -0.001998591236770153, -0.004831126891076565, 0.028066545724868774, -0.04946153610944748, -0.0024299416691064835, ...
[ "ACT2FN", "Linear", "ModelMLP", "Module", "None", "__init__", "act_fn", "class", "config", "def", "down_proj", "else", "forward", "gate_proj", "hidden_act", "hidden_size", "if", "intermediate_size", "is", "nn", "return", "self", "super", "up_proj", "x" ]
deepseek_v3/modeling_deepseek_v3.py:DeepseekV3TopkRouter
[ -0.00040611581061966717, 0.03441784903407097, 0.0062308176420629025, -0.007833028212189674, -0.0017876513302326202, 0.06978515535593033, 0.07168407738208771, -0.02883978560566902, -0.006349499803036451, 0.013470434583723545, 0.019463887438178062, -0.0206507109105587, 0.00022252919734455645, ...
[ "F", "ModelTopkRouter", "Module", "Parameter", "__init__", "class", "config", "def", "e_score_correction_bias", "empty", "float32", "forward", "hidden_size", "hidden_states", "linear", "n_routed_experts", "nn", "register_buffer", "return", "router_logits", "self", "super", ...
deepseek_v3/modeling_deepseek_v3.py:DeepseekV3NaiveMoe
[ -0.00035851384745910764, 0.041463855654001236, 0.00232943007722497, -0.0018635441083461046, -0.0012811865890398622, 0.054275721311569214, 0.05334394797682762, -0.01584012433886528, -0.005095628555864096, -0.012753630056977272, 0.014093051664531231, -0.030515534803271294, -0.00012921057350467...
[ "ACT2FN", "ModelNaiveMoe", "Module", "None", "Parameter", "__init__", "act_fn", "chunk", "class", "config", "continue", "current_hidden_states", "current_state", "def", "dim", "down_proj", "dtype", "empty", "expert_hit", "expert_idx", "expert_mask", "final_hidden_states", ...
deepseek_v3/modeling_deepseek_v3.py:DeepseekV3MoE
[ -0.0004175463109277189, 0.03760134056210518, -0.012415536679327488, -0.012297293171286583, -0.001625844044610858, 0.06952700763940811, 0.053445927798748016, -0.025777019560337067, -0.007922294549643993, -0.020810803398489952, 0.030743233859539032, -0.02885134145617485, -0.0007907514227554202...
[ "False", "ModelMLP", "ModelMoE", "ModelNaiveMoe", "ModelTopkRouter", "Module", "True", "__init__", "bool", "class", "config", "def", "denominator", "dim", "e_score_correction_bias", "expand", "experts", "forward", "gate", "gather", "group_idx", "group_mask", "group_scores...
deepseek_v3/modeling_deepseek_v3.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
deepseek_v3/modeling_deepseek_v3.py:apply_rotary_pos_emb
[ -0.0001444592053303495, 0.027112245559692383, 0.028019767254590988, 0.0028360087890177965, -0.0005707467789761722, 0.021667107939720154, 0.046510547399520874, -0.002183726755902171, 0.01293220091611147, 0.03652779385447502, 0.007884104736149311, 0.0017654155381023884, -0.0007834474672563374,...
[ "Model_rotary_pos_emb", "cos", "def", "k", "k_embed", "q", "q_embed", "return", "rotate_half", "sin", "unsqueeze", "unsqueeze_dim" ]
deepseek_v3/modeling_deepseek_v3.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
deepseek_v3/modeling_deepseek_v3.py:eager_attention_forward
[ 0, 0.021594731137156487, 0.01775064319372177, -0.018768195062875748, -0.00008832923776935786, 0.039797618985176086, 0.05359111353754997, -0.035049039870500565, 0.02069023996591568, 0.011645326390862465, 0.029395969584584236, 0.022499222308397293, 0.002741739386692643, -0.014584923163056374...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "causal_mask", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "key_states", "kwargs", "matmul", "module", "nn", "not", "num_key_value_groups", ...
deepseek_v3/modeling_deepseek_v3.py:apply_rotary_pos_emb_interleave
[ -0.00020252213289495558, 0.023753080517053604, 0.01792900823056698, -0.005881171673536301, -0.00091358000645414, 0.02752159908413887, 0.04841974005103111, -0.026493821293115616, 0.01587345264852047, 0.035401225090026855, 0.011248454451560974, 0.014217589050531387, -0.0020698297303169966, -...
[ "Model_rotary_pos_emb_interleave", "None", "b", "cos", "d", "def", "h", "k", "k_embed", "position_ids", "q", "q_embed", "r", "reshape", "return", "rotate_half", "s", "shape", "sin", "transpose", "unsqueeze", "unsqueeze_dim", "view" ]
deepseek_v3/modeling_deepseek_v3.py:yarn_get_mscale
[ -0.0001129502707044594, -0.010686078108847141, 0.0036001226399093866, 0.02400081790983677, -0.0002892955671995878, 0.049144528806209564, 0.048230212181806564, -0.01622912473976612, 0.015543386340141296, 0.06400217860937119, 0.016800571233034134, -0.006285928189754486, -0.0004464437661226839,...
[ "Model_get_mscale", "def", "if", "log", "math", "mscale", "return", "scale" ]
deepseek_v3/modeling_deepseek_v3.py:DeepseekV3Attention
[ -0.00008667633665027097, 0.022189142182469368, 0.025342930108308792, -0.006729955784976482, -0.00031678660889156163, 0.05136167258024216, 0.023315494880080223, -0.017796369269490242, 0.0034776132088154554, 0.014980487525463104, 0.029510432854294777, 0.031763140112161636, 0.001548734609968960...
[ "ALL_ATTENTION_FUNCTIONS", "F", "Linear", "ModelAttention", "ModelRMSNorm", "Module", "None", "Tensor", "True", "__init__", "_attn_implementation", "and", "apply_rotary_pos_emb", "apply_rotary_pos_emb_interleave", "attention_dropout", "attention_interface", "attention_mask", "attn_...
deepseek_v3/modeling_deepseek_v3.py:DeepseekV3DecoderLayer
[ -0.0002400130615569651, 0.04778926819562912, 0.0105819096788764, -0.0029299373272806406, -0.0007822648040018976, 0.035500600934028625, 0.041872501373291016, -0.03686600551009178, 0.004835818894207478, -0.007139944471418858, -0.0037548711989074945, 0.015929756686091423, -0.0013085156679153442...
[ "False", "GradientCheckpointingLayer", "ModelAttention", "ModelDecoderLayer", "ModelMLP", "ModelMoE", "ModelRMSNorm", "None", "Tensor", "_", "__init__", "attention_mask", "cache_position", "class", "config", "def", "else", "eps", "first_k_dense_replace", "forward", "hidden_si...
deepseek_v3/modeling_deepseek_v3.py:DeepseekV3PreTrainedModel
[ -0.00037101644556969404, 0.05074922740459442, 0.0014767908724024892, -0.005674369167536497, -0.0014840657822787762, 0.041670236736536026, 0.03398801386356354, -0.018972761929035187, -0.005121481604874134, -0.0031136281322687864, 0.008613401092588902, -0.009660976938903332, -0.002226098673418...
[ "ModelAttention", "ModelConfig", "ModelDecoderLayer", "ModelNaiveMoe", "ModelPreTrainedModel", "ModelTopkRouter", "PreTrainedModel", "True", "_can_compile_fullgraph", "_can_record_outputs", "_init_weights", "_keep_in_fp32_modules_strict", "_no_split_modules", "_skip_keys_device_placement",...
deepseek_v3/modeling_deepseek_v3.py:DeepseekV3Model
[ -0.00006877470877952874, 0.04306260496377945, 0.0003662143717519939, 0.001808292930945754, -0.00039424991700798273, 0.03924977034330368, 0.027474837377667427, -0.014298129826784134, 0.010148868896067142, 0.011270291171967983, 0.010933863930404186, 0, -0.0011354397283867002, 0.0106535088270...
[ "BaseModelOutputWithPast", "DynamicCache", "Embedding", "False", "ModelDecoderLayer", "ModelModel", "ModelPreTrainedModel", "ModelRMSNorm", "ModelRotaryEmbedding", "ModuleList", "None", "ValueError", "You", "__init__", "_keys_to_ignore_on_load_unexpected", "and", "arange", "attenti...
deepseek_v3/modeling_deepseek_v3.py:DeepseekV3ForCausalLM
[ -0.00028262686100788414, 0.03526614233851433, 0.008361488580703735, -0.001301150070503354, -0.0012087186332792044, 0.02741658128798008, 0.0389065183699131, -0.007394513580948114, 0.003640376031398773, 0.026734011247754097, 0.025823917239904404, 0.0026734010316431522, 0.0009385344455949962, ...
[ "CausalLMOutputWithPast", "GenerationMixin", "Linear", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "_pp_plan", "_tied_weights_keys", "_tp_plan", "attention_mask", "attentions", "auto_docstring", "cache_position", "can_return_tuple", "class", "colwi...
deepseek_v3/modeling_deepseek_v3.py:DeepseekV3ForSequenceClassification
[ -0.0002121782599715516, 0.015340753830969334, -0.017272552475333214, 0.0278406273573637, -0.0007244244916364551, 0.027954263612627983, 0.015227118507027626, -0.011079433374106884, 0.0032101948745548725, -0.013863496482372284, 0.02909061498939991, 0.008409005589783192, -0.003437465289607644, ...
[ "GenericForSequenceClassification", "ModelForSequenceClassification", "ModelPreTrainedModel", "class", "pass" ]
deepseek_v3/modeling_deepseek_v3.py:DeepseekV3ForTokenClassification
[ -0.00016310509818140417, 0.024394849315285683, -0.0069496952928602695, -0.024848707020282745, -0.0007446102099493146, 0.035854753106832504, 0.0356278270483017, -0.007885776460170746, 0.0015317696379497647, -0.018381234258413315, 0.040166404098272324, 0.024394849315285683, -0.0042832815088331...
[ "GenericForTokenClassification", "ModelForTokenClassification", "ModelPreTrainedModel", "class", "pass" ]
olmo/modeling_olmo.py:OlmoLayerNorm
[ -0.00006580990157090127, 0.04200602322816849, 0.03886118531227112, 0.037288766354322433, -0.0003738002560567111, 0.014601023867726326, 0.019655223935842514, -0.03953507915139198, 0.005728093907237053, 0.05233905464410782, 0.00814287830144167, 0.0017338715260848403, 0.0005089298938401043, 0...
[ "F", "ModelLayerNorm", "Module", "None", "__init__", "class", "def", "dtype", "eps", "float32", "forward", "hidden_size", "hidden_states", "layer_norm", "nn", "normalized_shape", "orig_dtype", "return", "self", "super", "to", "torch" ]
olmo/modeling_olmo.py:OlmoMLP
[ -0.0002434849739074707, 0.024471141397953033, 0.0228551235049963, 0.026548879221081734, -0.0009450823999941349, 0.060485273599624634, 0.036244992166757584, -0.005078915972262621, 0, -0.005021201446652412, 0.028164898976683617, -0.04640282690525055, -0.0015583038330078125, 0.009176678024232...
[ "ACT2FN", "Linear", "ModelMLP", "Module", "__init__", "act_fn", "class", "config", "def", "down_proj", "forward", "gate_proj", "hidden_act", "hidden_size", "intermediate_size", "nn", "return", "self", "super", "up_proj", "x" ]
olmo/modeling_olmo.py:OlmoRotaryEmbedding
[ -0.00035672387457452714, 0.049173012375831604, 0.0012659123167395592, -0.005414885468780994, -0.0017269093077629805, 0.04285076633095741, 0.043319083750247955, 0.004683143924921751, -0.0014342128997668624, 0.02072291262447834, 0.0032489311415702105, -0.005883199628442526, -0.0014415303012356...
[ "False", "ModelRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "arange", "attention_factor", "attention_scaling", "base", "cat", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cpu", "def", "default", "device", ...
olmo/modeling_olmo.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
olmo/modeling_olmo.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
olmo/modeling_olmo.py:eager_attention_forward
[ 0, 0.021594731137156487, 0.01775064319372177, -0.018768195062875748, -0.00008832923776935786, 0.039797618985176086, 0.05359111353754997, -0.035049039870500565, 0.02069023996591568, 0.011645326390862465, 0.029395969584584236, 0.022499222308397293, 0.002741739386692643, -0.014584923163056374...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "causal_mask", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "key_states", "kwargs", "matmul", "module", "nn", "not", "num_key_value_groups", ...
olmo/modeling_olmo.py:apply_rotary_pos_emb
[ -0.00016319907444994897, 0.027587739750742912, 0.0264524407684803, 0.0003601023054216057, -0.0006527962977997959, 0.023273607715964317, 0.045638974756002426, -0.004995310679078102, 0.013112691231071949, 0.03632953390479088, 0.006130608730018139, 0.0023841257207095623, -0.0006918221479281783,...
[ "Model_rotary_pos_emb", "cos", "def", "k", "k_embed", "q", "q_embed", "return", "rotate_half", "sin", "unsqueeze", "unsqueeze_dim" ]
olmo/modeling_olmo.py:OlmoAttention
[ -0.00009394854714628309, 0.043381400406360626, 0.03191792219877243, -0.008766190148890018, -0.00040213490137830377, 0.023713666945695877, 0.031693149358034134, -0.006967997178435326, 0, 0.013205477967858315, 0.01618373580276966, 0.0334913395345211, 0.0007656367961317301, -0.015958961099386...
[ "ALL_ATTENTION_FUNCTIONS", "Linear", "ModelAttention", "Module", "None", "Tensor", "True", "__init__", "_attn_implementation", "apply_rotary_pos_emb", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "cache_kwargs", "cache_position", "cla...
olmo/modeling_olmo.py:OlmoDecoderLayer
[ -0.00016472279094159603, 0.041267022490501404, 0.015785200521349907, -0.001987244002521038, -0.0005602336605079472, 0.035629451274871826, 0.04464956372976303, -0.032923415303230286, 0.0061449529603123665, -0.00715971551835537, -0.00597582571208477, 0.020971765741705894, -0.001973150065168738...
[ "False", "GradientCheckpointingLayer", "ModelAttention", "ModelDecoderLayer", "ModelLayerNorm", "ModelMLP", "None", "Tensor", "_", "__init__", "attention_mask", "cache_position", "class", "config", "def", "forward", "hidden_size", "hidden_states", "input_layernorm", "kwargs", ...
olmo/modeling_olmo.py:OlmoPreTrainedModel
[ -0.00034975787275470793, 0.027630871161818504, 0.016904963180422783, 0.011250545270740986, -0.0018799485405907035, 0.029612833634018898, 0.03077869303524494, -0.03241089731454849, -0.00408050836995244, 0.003993069287389517, 0.006383080966770649, 0.004459412768483162, -0.004255387466400862, ...
[ "ModelAttention", "ModelConfig", "ModelDecoderLayer", "ModelPreTrainedModel", "PreTrainedModel", "True", "_can_compile_fullgraph", "_can_record_outputs", "_no_split_modules", "_skip_keys_device_placement", "_supports_attention_backend", "_supports_flash_attn", "_supports_flex_attn", "_supp...
olmo/modeling_olmo.py:OlmoModel
[ -0.00008881748362910002, 0.046600159257650375, -0.008836020715534687, -0.009173702448606491, -0.0007597851799800992, 0.03962139040231705, 0.0398465134203434, -0.015195704065263271, 0.011368637904524803, 0.006584804970771074, 0.01553338672965765, -0.003517524106428027, -0.0019698136020451784,...
[ "BaseModelOutputWithPast", "DynamicCache", "Embedding", "False", "ModelDecoderLayer", "ModelLayerNorm", "ModelModel", "ModelPreTrainedModel", "ModelRotaryEmbedding", "ModuleList", "None", "ValueError", "You", "__init__", "and", "arange", "attention_mask", "auto_docstring", "cache...
olmo/modeling_olmo.py:OlmoForCausalLM
[ -0.00028262686100788414, 0.03526614233851433, 0.008361488580703735, -0.001301150070503354, -0.0012087186332792044, 0.02741658128798008, 0.0389065183699131, -0.007394513580948114, 0.003640376031398773, 0.026734011247754097, 0.025823917239904404, 0.0026734010316431522, 0.0009385344455949962, ...
[ "CausalLMOutputWithPast", "GenerationMixin", "Linear", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "_pp_plan", "_tied_weights_keys", "_tp_plan", "attention_mask", "attentions", "auto_docstring", "cache_position", "can_return_tuple", "class", "colwi...
glm_image/modeling_glm_image.py:GlmImageVisionMLP
[ -0.0001662031572777778, 0.03708730265498161, 0.032536715269088745, 0.034129418432712555, -0.00047283468302339315, 0.04709859937429428, 0.03162659704685211, -0.03481200709939003, 0.003341838950291276, 0, 0.049373894929885864, -0.020136358216404915, 0.0019482210045680404, -0.0018913387320935...
[ "ACT2FN", "Linear", "ModelVisionMLP", "Module", "__init__", "activation_fn", "class", "config", "def", "fc1", "fc2", "forward", "hidden_act", "hidden_size", "hidden_states", "intermediate_size", "nn", "return", "self", "super" ]
glm_image/modeling_glm_image.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
glm_image/modeling_glm_image.py:eager_attention_forward
[ 0, 0.021594731137156487, 0.01775064319372177, -0.018768195062875748, -0.00008832923776935786, 0.039797618985176086, 0.05359111353754997, -0.035049039870500565, 0.02069023996591568, 0.011645326390862465, 0.029395969584584236, 0.022499222308397293, 0.002741739386692643, -0.014584923163056374...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "causal_mask", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "key_states", "kwargs", "matmul", "module", "nn", "not", "num_key_value_groups", ...
glm_image/modeling_glm_image.py:GlmImageVisionAttention
[ -0.00009369172039441764, 0.03032076172530651, 0.03032076172530651, 0.022627433761954308, -0.00019975782197434455, 0.01708371192216873, 0.029415663331747055, -0.04299212619662285, 0.0049497513100504875, 0.03507252410054207, 0.04005055874586105, 0.02036469057202339, 0.0009263106039725244, -0...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "ModelVisionAttention", "Module", "None", "_", "__init__", "_attn_implementation", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_outputs", "cat", "class", "config", "contiguous", "cu_seq_lens_k", ...
glm_image/modeling_glm_image.py:GlmImageVisionPatchEmbed
[ -0.000021998759621055797, -0.0018352000042796135, 0.024992341175675392, 0.016586845740675926, 0, 0.0038945465348660946, 0.02510441467165947, -0.0187162384390831, 0.010590924881398678, 0.0013238656101748347, 0.015802333131432533, -0.008909826166927814, -0.0014359388733282685, -0.01658684574...
[ "Conv2d", "ModelVisionPatchEmbed", "Module", "__init__", "class", "config", "def", "dtype", "embed_dim", "forward", "hidden_size", "hidden_states", "in_channels", "kernel_size", "nn", "patch_size", "proj", "return", "self", "stride", "super", "target_dtype", "view", "we...
glm_image/modeling_glm_image.py:GlmImageVisionEmbeddings
[ -0.0002974536910187453, 0.028899597004055977, 0.005017291288822889, 0.033716198056936264, -0.0012686578556895256, 0.01662873663008213, 0.02431236021220684, -0.009690539911389351, 0.005217982921749353, 0.014793841168284416, 0.023280231282114983, 0.04426684230566025, 0.0002795347827486694, 0...
[ "Embedding", "F", "False", "ModelVisionEmbeddings", "Module", "__init__", "adapted_pos_embed", "adapted_pos_embed_fp32", "align_corners", "bilinear", "border", "cat", "class", "config", "def", "device", "dim", "dtype", "embed_dim", "embeddings", "float32", "for", "forward...
glm_image/modeling_glm_image.py:GlmImageVisionBlock
[ -0.0001120106753660366, 0.014563151635229588, 0.015353400260210037, 0.017385467886924744, -0.0000956941585172899, 0.04131871089339256, 0.02980365976691246, -0.029352089390158653, 0.009652321226894855, 0.027207128703594208, 0.02686845138669014, 0.02088513970375061, 0.002046179259195924, -0....
[ "GradientCheckpointingLayer", "LayerNorm", "ModelVisionAttention", "ModelVisionBlock", "ModelVisionMLP", "__init__", "attn", "class", "config", "cu_seqlens", "def", "eps", "forward", "hidden_size", "hidden_states", "kwargs", "layer_norm_eps", "mlp", "nn", "norm1", "norm2", ...
glm_image/modeling_glm_image.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
glm_image/modeling_glm_image.py:apply_rotary_pos_emb
[ -0.00020313466666266322, 0.028396088629961014, 0.021097494289278984, -0.006728390231728554, -0.0008695589494891465, 0.02337830513715744, 0.04196690768003464, -0.0029650533106178045, 0.011575112119317055, 0.03329982981085777, 0.0030790939927101135, 0.00803985632956028, -0.0009693444008007646,...
[ "Model_rotary_pos_emb", "cat", "cos", "def", "dim", "k", "k_embed", "k_pass", "k_rot", "q", "q_embed", "q_pass", "q_rot", "return", "rotary_dim", "rotate_half", "shape", "sin", "torch", "unsqueeze", "unsqueeze_dim" ]
glm_image/modeling_glm_image.py:GlmImageTextAttention
[ -0.00006889815267641097, 0.037298038601875305, 0.039544906467199326, -0.010841147042810917, -0.00039144683978520334, 0.022581040859222412, 0.03684866428375244, -0.02179463766515255, -0.0013832291588187218, 0.012301611714065075, 0.02370447665452957, 0.03639928996562958, -0.0026822008658200502...
[ "ALL_ATTENTION_FUNCTIONS", "Linear", "ModelTextAttention", "Module", "None", "Tensor", "True", "__init__", "_attn_implementation", "apply_rotary_pos_emb", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "cache_kwargs", "cache_position", ...
glm_image/modeling_glm_image.py:GlmImageRMSNorm
[ -0.00009884802420856431, 0.04179859161376953, 0.03230918198823929, 0.053095508366823196, -0.00046952810953371227, 0.03931327164173126, 0.022028988227248192, -0.02937198430299759, 0.007964326068758965, 0.042476408183574677, 0.019995542243123055, 0.006975846365094185, 0.0028524715453386307, ...
[ "ModelRMSNorm", "Module", "Parameter", "True", "__init__", "class", "def", "eps", "extra_repr", "f", "float32", "forward", "hidden_size", "hidden_states", "keepdim", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "shape", "super", "to", "torch", "tuple", ...
glm_image/modeling_glm_image.py:GlmImageTextMLP
[ -0.00034343745210208, 0.012919093482196331, 0.03180084750056267, 0.016368083655834198, -0.0015272005693987012, 0.04536297172307968, 0.024084463715553284, -0.027825741097331047, -0.005085797049105167, -0.0075410096906125546, 0.04325850307941437, -0.03507446125149727, -0.0005151561927050352, ...
[ "ACT2FN", "Linear", "ModelTextMLP", "Module", "__init__", "activation_fn", "chunk", "class", "config", "def", "dim", "down_proj", "forward", "gate", "gate_up_proj", "hidden_act", "hidden_size", "hidden_states", "intermediate_size", "nn", "return", "self", "super", "up_s...
glm_image/modeling_glm_image.py:GlmImageTextDecoderLayer
[ -0.0001190652183140628, 0.04515658691525459, 0.01693372055888176, 0.005249453242868185, -0.00046214944450184703, 0.03815731406211853, 0.03251274302601814, -0.03747996687889099, 0.008918425999581814, -0.005221230443567038, 0.014111433178186417, 0.03702840209007263, -0.002610615221783519, 0....
[ "False", "GradientCheckpointingLayer", "ModelRMSNorm", "ModelTextAttention", "ModelTextDecoderLayer", "ModelTextMLP", "None", "Tensor", "_", "__init__", "attention_mask", "auto_docstring", "cache_position", "class", "config", "def", "eps", "forward", "hidden_size", "hidden_stat...
glm_image/modeling_glm_image.py:GlmImagePreTrainedModel
[ -0.00023248510842677206, 0.04270627349615097, 0.007893845438957214, 0.007723474875092506, -0.0011854965705424547, 0.025442035868763924, 0.01885436475276947, -0.0324840284883976, -0.003946922719478607, 0.011017309501767159, 0.006758040748536587, -0.0013771636877208948, -0.00477038137614727, ...
[ "ModelConfig", "ModelPreTrainedModel", "ModelTextAttention", "ModelTextDecoderLayer", "ModelVisionBlock", "PreTrainedModel", "True", "_can_compile_fullgraph", "_can_record_outputs", "_init_weights", "_no_split_modules", "_skip_keys_device_placement", "_supports_attention_backend", "_suppor...
glm_image/modeling_glm_image.py:GlmImageModelOutputWithPast
[ -0.0002648343506734818, 0.008071142248809338, 0.014355102553963661, 0.0052174171432852745, -0.0014196563279256225, 0.042200542986392975, 0.04911866411566734, -0.03851087763905525, 0.016949398443102837, 0.0025078190956264734, 0.008071142248809338, 0.03182335942983627, -0.001830419758334756, ...
[ "ModelModelOutputWithPast", "ModelOutput", "None", "attentions", "class", "hidden_states", "last_hidden_state", "past_key_values", "r", "rope_deltas" ]
glm_image/modeling_glm_image.py:GlmImageVQVAEVectorQuantizer
[ -0.00016392211546190083, 0.026479996740818024, 0.02962169237434864, 0.007573727983981371, -0.0005014088237658143, 0.018064742907881737, 0.026592200621962547, -0.030070506036281586, 0.0035344064235687256, 0.02490914985537529, 0.05161355435848236, 0.017167117446660995, 0.00023141947167459875, ...
[ "Embedding", "F", "ModelVQVAEVectorQuantizer", "Module", "True", "__init__", "argmin", "bd", "beta", "bn", "class", "config", "contiguous", "def", "detach", "dim", "distances", "dn", "einsum", "embed_dim", "embedding", "embedding_dim", "forward", "getattr", "hidden_st...
glm_image/modeling_glm_image.py:GlmImageVQVAEModelOutput
[ -0.0001690711360424757, 0.02130119316279888, 0.016315806657075882, 0.0062317317351698875, -0.0008851892198435962, 0.043962035328149796, 0.04622812196612358, -0.020847976207733154, 0.012803376652300358, -0.00753473024815321, 0.025833360850811005, 0.02821275033056736, -0.003908995538949966, ...
[ "BaseModelOutputWithPooling", "ModelVQVAEModelOutput", "None", "class", "embedding_loss", "image_tokens", "quantized_last_hidden_state", "r" ]
glm_image/modeling_glm_image.py:GlmImageVQVAE
[ -0.00011358158371876925, 0.018595682457089424, 0.003972713835537434, 0.01679246500134468, -0.0004419996403157711, 0.040797799825668335, 0.03155630826950073, -0.03245791792869568, 0.005606879945844412, 0.011101058684289455, 0.041474007070064545, 0.013749535195529461, -0.002296285005286336, ...
[ "Conv2d", "ModelPreTrainedModel", "ModelVQVAE", "ModelVQVAEConfig", "ModelVQVAEModelOutput", "ModelVQVAEVectorQuantizer", "__init__", "_can_record_outputs", "_no_split_modules", "check_model_inputs", "class", "config", "conv_hidden_states", "def", "emb_loss", "embed_dim", "embedding_...
glm_image/modeling_glm_image.py:GlmImageVisionModel
[ -0.00009591030539013445, 0.026338715106248856, -0.019865641370415688, 0.015401451848447323, -0.0002048992901109159, 0.02020045556128025, 0.03549030050635338, -0.028235994279384613, 0.005775544326752424, 0.02790118008852005, 0.02566908672451973, 0.015178241766989231, -0.0005754618323408067, ...
[ "BaseModelOutputWithPooling", "F", "False", "ModelPreTrainedModel", "ModelVisionAttention", "ModelVisionBlock", "ModelVisionConfig", "ModelVisionEmbeddings", "ModelVisionModel", "ModelVisionPatchEmbed", "ModuleList", "_", "__init__", "_can_record_outputs", "_no_split_modules", "append"...
glm_image/modeling_glm_image.py:GlmImageTextRotaryEmbedding
[ -0.0003361947601661086, 0.04279905557632446, 0.003975868225097656, -0.020347092300653458, -0.001600579358637333, 0.05449278652667999, 0.028766578063368797, -0.010582826100289822, -0.008477955125272274, 0.02981901355087757, 0.014266351237893105, -0.0049406010657548904, -0.0008075982914306223,...
[ "False", "ModelTextRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "apply_mrope", "arange", "attention_factor", "attention_scaling", "base", "cat", "chunk", "chunks", "class", "clone", "compute_default_rope_parameters", "config", "cos",...
glm_image/modeling_glm_image.py:GlmImageTextModel
[ -0.000022241098122322, 0.053811248391866684, 0.009210431016981602, -0.011275800876319408, -0.00037853477988392115, 0.02143518626689911, 0.03751157596707344, -0.024672791361808777, 0.008261477574706078, 0.011275800876319408, 0.015294898301362991, 0.009880281053483486, -0.0032236508559435606, ...
[ "BaseModelOutputWithPast", "DynamicCache", "Embedding", "False", "ModelPreTrainedModel", "ModelRMSNorm", "ModelTextConfig", "ModelTextDecoderLayer", "ModelTextModel", "ModelTextRotaryEmbedding", "ModuleList", "None", "ValueError", "You", "__init__", "and", "arange", "attention_mask...
glm_image/modeling_glm_image.py:GlmImageModel
[ -0.000048940139095066115, 0.04362663999199867, -0.017674382776021957, -0.0018177765887230635, -0.0003146151721011847, 0.04206055402755737, 0.034006405621767044, -0.036914847791194916, 0.004921979736536741, 0.014765939675271511, 0.0422842800617218, 0.0234912671148777, -0.0006187431863509119, ...
[ "Ensure", "F", "False", "ModelConfig", "ModelModel", "ModelModelOutputWithPast", "ModelPreTrainedModel", "ModelTextDecoderLayer", "ModelTextModel", "ModelVQVAE", "ModelVisionBlock", "ModelVisionModel", "None", "Number", "True", "VQVAE", "ValueError", "You", "_", "__init__", "...
glm_image/modeling_glm_image.py:GlmImageCausalLMOutputWithPast
[ -0.0002988479973282665, 0.017503436654806137, 0.020285440608859062, -0.0002644352207425982, -0.0015503872418776155, 0.049844224005937576, 0.049612391740083694, -0.02874736674129963, 0.016923854127526283, 0.0007607040461152792, 0.017155686393380165, 0.02156052552163601, -0.0010577408829703927...
[ "ModelCausalLMOutputWithPast", "ModelOutput", "None", "attentions", "class", "hidden_states", "logits", "loss", "past_key_values", "r", "rope_deltas" ]
glm_image/modeling_glm_image.py:GlmImageForConditionalGeneration
[ -0.0002754932502284646, 0.03390686213970184, -0.005340330768376589, 0.025430146604776382, -0.0009253747994080186, 0.020344117656350136, 0.02701246738433838, -0.04475705698132515, -0.005792422220110893, 0.010906707495450974, 0.03752359375357628, 0.009719966910779476, 0.0019920282065868378, ...
[ "False", "GenerationMixin", "Linear", "ModelCausalLMOutputWithPast", "ModelConfig", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "Tensor", "True", "__init__", "_checkpoint_conversion_mapping", "_expand_dict_for_generation", "_expand_dict_for_generation_v...
pix2struct/modeling_pix2struct.py:Pix2StructLayerNorm
[ -0.00005301048804540187, 0.037010956555604935, 0.041497133672237396, 0.04912363365292549, -0.00022606125276070088, 0.025571206584572792, 0.01772039756178856, -0.016598854213953018, 0.005327334627509117, 0.05360981076955795, 0.017383934929966927, 0.005691836588084698, 0.0024673971347510815, ...
[ "ModelLayerNorm", "Module", "Parameter", "True", "__init__", "bfloat16", "class", "def", "dtype", "eps", "float16", "float32", "forward", "hidden_size", "hidden_states", "if", "in", "keepdim", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "super", "to", ...
pix2struct/modeling_pix2struct.py:Pix2StructVisionEmbeddings
[ -0.00017313203716184944, -0.0022871186956763268, 0.01414888259023428, 0.04136701673269272, -0.000593088218010962, 0.02693401835858822, 0.007955195382237434, -0.00602321932092309, 0.013921591453254223, 0.004176477435976267, 0.006705093197524548, 0.02829776518046856, -0.001917770248837769, -...
[ "Dropout", "Embedding", "Linear", "ModelVisionEmbeddings", "Module", "__init__", "class", "col_embeddings", "col_indices", "column_embedder", "config", "def", "dropout", "dropout_rate", "embeddings", "flattened_patches", "forward", "hidden_size", "long", "nn", "patch_embed_hi...
pix2struct/modeling_pix2struct.py:Pix2StructVisionAttention
[ -0.00013926601968705654, 0.04462118446826935, 0.02914951741695404, 0.0021161427721381187, -0.00039590088999830186, 0.014070247299969196, 0.03879128023982048, -0.015247439965605736, 0.005129193887114525, 0.033409830182790756, 0.022422704845666885, 0.03520364686846733, 0.00017955682415049523, ...
[ "False", "Linear", "ModelVisionAttention", "Module", "None", "True", "__init__", "and", "attention_dropout", "attention_mask", "attn_output", "attn_weights", "batch_size", "class", "config", "contiguous", "d_kv", "def", "device", "dim", "dropout", "dtype", "elif", "finf...
pix2struct/modeling_pix2struct.py:Pix2StructVisionMlp
[ -0.00036044305306859314, 0.05755438283085823, 0.022951848804950714, 0.02702958881855011, -0.0012378852115944028, 0.04194246605038643, 0.03565109521150589, -0.04543767124414444, -0.004893287550657988, 0.011068150401115417, 0.04217547923326492, -0.002534023951739073, 0.001500025624409318, 0....
[ "ACT2FN", "Dropout", "Linear", "ModelVisionMlp", "Module", "Tensor", "__init__", "act", "and", "class", "config", "d_ff", "def", "dense_act_fn", "dropout", "dropout_rate", "dtype", "forward", "hidden_gelu", "hidden_linear", "hidden_size", "hidden_states", "if", "int8", ...
pix2struct/modeling_pix2struct.py:Pix2StructVisionLayer
[ -0.00008012827311176807, 0.03721454739570618, 0.01703798584640026, 0.0367661789059639, 0, 0.03362760320305824, 0.02981647476553917, -0.020288653671741486, 0.008070625364780426, 0.025669071823358536, 0.030489027500152588, 0.02858346328139305, 0.0025921277701854706, -0.011265247128903866, ...
[ "False", "GradientCheckpointingLayer", "ModelLayerNorm", "ModelVisionAttention", "ModelVisionLayer", "ModelVisionMlp", "None", "__init__", "attention", "attention_mask", "attention_output", "chunk_size_feed_forward", "class", "config", "def", "eps", "forward", "hidden_size", "hid...
pix2struct/modeling_pix2struct.py:Pix2StructVisionEncoder
[ -0.000022785628971178085, 0.025127289816737175, 0.009871435351669788, 0.05070328339934349, 0.00019455421715974808, 0.03477437421679497, 0.018733292818069458, -0.03656918182969093, 0.007291401270776987, 0.02523946575820446, 0.02916560508310795, 0.00678661186248064, 0.0007396565633825958, -0...
[ "BaseModelOutput", "False", "ModelVisionEncoder", "ModelVisionLayer", "Module", "ModuleList", "None", "True", "_", "__init__", "all_hidden_states", "all_self_attentions", "attention_mask", "attentions", "class", "config", "def", "else", "enumerate", "for", "forward", "gradi...
pix2struct/modeling_pix2struct.py:Pix2StructPreTrainedModel
[ -0.000251176068559289, 0.05335621163249016, -0.0030497408006340265, 0.007125562522560358, -0.001225596759468317, 0.037166934460401535, 0.01846945844590664, -0.03602684289216995, -0.003591283457353711, 0.015049188397824764, 0.017557386308908463, 0.0004738499119412154, -0.0031352476216852665, ...
[ "Conv2d", "DUMMY_INPUTS", "DUMMY_MASK", "Embedding", "False", "In", "Linear", "Model", "ModelConfig", "ModelLayerNorm", "ModelPreTrainedModel", "ModelTextAttention", "ModelTextDenseGatedActDense", "ModelTextModel", "None", "PreTrainedModel", "See", "ValueError", "_can_compile_ful...
pix2struct/modeling_pix2struct.py:Pix2StructVisionModel
[ 0.000033066618925658986, 0.03475717455148697, 0.0015276339836418629, 0.04170861095190048, 0.00037840474396944046, 0.011604412458837032, 0.012669551186263561, -0.01009079348295927, 0.003097312757745385, 0.0169301088899374, 0.014911949634552002, 0.012893791310489178, -0.003643897594884038, 0...
[ "BaseModelOutput", "ModelLayerNorm", "ModelPreTrainedModel", "ModelVisionConfig", "ModelVisionEmbeddings", "ModelVisionEncoder", "ModelVisionLayer", "ModelVisionModel", "None", "True", "__init__", "_no_split_modules", "attention_mask", "attentions", "auto_docstring", "class", "config...
pix2struct/modeling_pix2struct.py:Pix2StructTextDenseGatedActDense
[ -0.00033420242834836245, 0.03045601397752762, 0.031153479591012, -0.0066549875773489475, -0.001227830653078854, 0.02406257577240467, 0.03347836807370186, -0.037663161754608154, -0.001620155293494463, -0.004330101422965527, 0.028479861095547676, -0.012089409865438938, -0.00016165226406883448,...
[ "ACT2FN", "Dropout", "Linear", "ModelTextDenseGatedActDense", "Module", "Tensor", "__init__", "act", "and", "class", "config", "d_ff", "def", "dense_act_fn", "dropout", "dropout_rate", "dtype", "forward", "hidden_gelu", "hidden_linear", "hidden_size", "hidden_states", "if...
pix2struct/modeling_pix2struct.py:Pix2StructTextLayerFF
[ -0.00012050191435264423, 0.0342508964240551, 0.03175579756498337, 0.021435163915157318, -0.0002817618369590491, 0.0408288836479187, 0.025404639542102814, -0.03447772562503815, 0.004848429933190346, -0.005415497813373804, 0.03311676159501076, -0.0030479896813631058, 0.0017295569414272904, 0...
[ "DenseReluDense", "Dropout", "ModelLayerNorm", "ModelTextDenseGatedActDense", "ModelTextLayerFF", "Module", "__init__", "class", "config", "def", "dropout", "dropout_rate", "eps", "forward", "forwarded_states", "hidden_size", "hidden_states", "layer_norm", "layer_norm_epsilon", ...
pix2struct/modeling_pix2struct.py:Pix2StructTextAttention
[ -0.00011677078873617575, 0.03326474875211716, 0.05102093145251274, -0.03236570209264755, -0.0006953568663448095, 0.017980944365262985, 0.029219036921858788, -0.04023236408829689, 0.004888569470494986, 0.00831618718802929, 0.021914277225732803, 0.04068188741803169, 0.0018402374116703868, 0....
[ "Embedding", "EncoderDecoderCache", "False", "Instantiating", "Linear", "ModelTextAttention", "Module", "None", "Please", "True", "__class__", "__init__", "__name__", "_relative_position_bucket", "a", "abs", "and", "arange", "attn_output", "attn_weights", "batch_size", "bid...
pix2struct/modeling_pix2struct.py:Pix2StructTextLayerSelfAttention
[ 0.000052020906878169626, 0.05058355629444122, 0.04028778895735741, -0.003357315668836236, 0.00026753608835861087, 0.025851331651210785, 0.023053567856550217, -0.028649095445871353, 0.006462832912802696, 0.0025879309978336096, 0.025403689593076706, 0.040063969790935516, 0.0008498205570504069,...
[ "Dropout", "False", "ModelLayerNorm", "ModelTextAttention", "ModelTextLayerSelfAttention", "Module", "None", "__init__", "attention", "attention_mask", "attention_output", "cache_position", "class", "config", "def", "dropout", "dropout_rate", "eps", "forward", "has_relative_att...
pix2struct/modeling_pix2struct.py:Pix2StructTextLayerCrossAttention
[ 0, 0.04629536718130112, 0.05168899893760681, -0.0011587886838242412, -0.000184352757059969, 0.0312381349503994, 0.028766052797436714, -0.03168760612607002, 0.006039747502654791, 0.014270658604800701, 0.0193271916359663, 0.04764377325773239, 0.0008462669211439788, -0.016405640169978142, -...
[ "Dropout", "False", "ModelLayerNorm", "ModelTextAttention", "ModelTextLayerCrossAttention", "Module", "None", "__init__", "attention", "attention_mask", "attention_output", "cache_position", "class", "config", "def", "dropout", "dropout_rate", "eps", "forward", "has_relative_at...
pix2struct/modeling_pix2struct.py:Pix2StructTextBlock
[ -0.0001432684512110427, 0.02546994760632515, 0.04029911756515503, -0.015961166471242905, -0.0007110360311344266, 0.028979141265153885, 0.019130760803818703, -0.038261521607637405, 0.005037389695644379, -0.002363045234233141, 0.015055568888783455, 0.026035945862531662, 0.0019526960095390677, ...
[ "False", "GradientCheckpointingLayer", "ModelTextBlock", "ModelTextLayerCrossAttention", "ModelTextLayerFF", "ModelTextLayerSelfAttention", "None", "True", "__init__", "and", "any", "attention_mask", "attention_outputs", "cache_position", "clamp", "clamp_value", "class", "config", ...