identifier
stringlengths
24
102
embedding
listlengths
2.56k
2.56k
tokens
listlengths
4
448
gpt_neo/modeling_gpt_neo.py:GPTNeoPreTrainedModel
[ -0.00010462188220117241, 0.034660615026950836, 0.02351970225572586, -0.016655100509524345, -0.000597839301917702, 0.029709096997976303, 0.020593805238604546, -0.03286006301641464, 0.003038430353626609, 0.003221298800781369, 0.016655100509524345, 0.007877412252128124, 0.0034463678020983934, ...
[ "False", "Model", "ModelBlock", "ModelConfig", "ModelSelfAttention", "PreTrainedModel", "True", "_can_compile_fullgraph", "_init_weights", "_no_split_modules", "_skip_keys_device_placement", "_supports_flash_attn", "attention_type", "base_model_prefix", "bias", "bitwise_xor", "bool",...
gpt_neo/modeling_gpt_neo.py:GPTNeoModel
[ 0.000011221967724850401, 0.02231319434940815, 0.01355526503175497, -0.015507670119404793, 0, 0.02755679376423359, 0.029899679124355316, -0.012551171705126762, 0.007028656080365181, 0.006052454002201557, 0.017738988623023033, 0.010710333473980427, -0.0000631916627753526, 0.02287102304399013...
[ "AttentionMaskConverter", "BaseModelOutputWithPast", "BlockMask", "Dropout", "DynamicCache", "Embedding", "False", "LayerNorm", "Model", "ModelBlock", "ModelPreTrainedModel", "ModuleList", "None", "Setting", "Tensor", "True", "ValueError", "You", "__init__", "_attn_implementati...
gpt_neo/modeling_gpt_neo.py:GPTNeoForCausalLM
[ -0.00022742200235370547, 0.0442291721701622, 0.020309314131736755, -0.014667837880551815, -0.000923791725654155, 0.02674059569835663, 0.03678242489695549, -0.012749736197292805, 0.004372144117951393, 0.009082776494324207, 0.02019648440182209, 0.013370297849178314, 0.0006910808151587844, 0....
[ "CausalLMOutputWithPast", "GenerationMixin", "Linear", "Model", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "cache_position", "class", "config", "def", "else", "forward", "hidden_size", "hidde...
gpt_neo/modeling_gpt_neo.py:GPTNeoForSequenceClassification
[ -0.0003615770547185093, 0.038033612072467804, 0.009909359738230705, 0.0046110027469694614, -0.0013532290467992425, 0.024859318509697914, 0.020276956260204315, -0.012143261730670929, -0.0015393875073641539, -0.012257820926606655, 0.037117138504981995, 0.028639767318964005, 0.00216230237856507...
[ "BCEWithLogitsLoss", "Cannot", "CrossEntropyLoss", "Linear", "MSELoss", "Model", "ModelModel", "ModelPreTrainedModel", "None", "Results", "SequenceClassifierOutputWithPast", "ValueError", "__class__", "__init__", "__name__", "and", "arange", "argmax", "attention_mask", "attenti...
gpt_neo/modeling_gpt_neo.py:GPTNeoForTokenClassification
[ -0.00025373115204274654, 0.04019952937960625, 0.01873706839978695, -0.020326880738139153, -0.0008658797014504671, 0.03156912326812744, 0.038155484944581985, 0.007949059829115868, -0.0006103742052800953, 0.007721943315118551, 0.04496896639466286, 0.014421865344047546, 0.0007416756125167012, ...
[ "CrossEntropyLoss", "Dropout", "Linear", "Model", "ModelModel", "ModelPreTrainedModel", "None", "TokenClassifierOutput", "__init__", "attention_mask", "attentions", "auto_docstring", "class", "classifier", "classifier_dropout", "config", "def", "device", "dropout", "else", "f...
gpt_neo/modeling_gpt_neo.py:GPTNeoForQuestionAnswering
[ -0.00021967558132018894, 0.031167466193437576, 0.01863272488117218, 0.01010684110224247, -0.0007163718109950423, 0.04788045585155487, 0.03365182876586914, 0.023940227925777435, 0.003147801151499152, 0.02845725230872631, 0.017955170944333076, 0.02766677364706993, 0.0011786610120907426, 0.00...
[ "CrossEntropyLoss", "Linear", "Model", "ModelModel", "ModelPreTrainedModel", "None", "QuestionAnsweringModelOutput", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "clamp", "class", "config", "contiguous", "def", "dim", "else", "end_logits", "end_loss", ...
decision_transformer/modeling_decision_transformer.py:eager_attention_forward
[ -0.000011272472875134554, 0.03738809749484062, 0.017117442563176155, -0.01813097484409809, -0.00019267680181656033, 0.032658278942108154, 0.05292893573641777, -0.017680516466498375, 0.018806664273142815, 0.02117157354950905, 0.020270656794309616, 0.03423488512635231, 0.0022945257369428873, ...
[ "Model_attention_forward", "None", "attention_mask", "attn_dropout", "attn_output", "attn_weights", "bias", "causal_mask", "def", "device", "dim", "dtype", "finfo", "float", "full", "functional", "if", "is", "is_cross_attention", "key", "key_length", "kwargs", "layer_idx"...
decision_transformer/modeling_decision_transformer.py:DecisionTransformerGPT2Attention
[ -0.00009797888924367726, 0.044766247272491455, 0.03801756724715233, -0.014397185295820236, -0.0007346220081672072, 0.01422846782952547, 0.03801756724715233, -0.011247800663113594, -0.0019683651626110077, 0.02047099731862545, 0.01765904761850834, 0.03374340385198593, -0.0005342705408111215, ...
[ "ALL_ATTENTION_FUNCTIONS", "Conv1D", "Dropout", "EncoderDecoderCache", "False", "If", "ModelGPT2Attention", "Module", "None", "Please", "True", "ValueError", "_", "__init__", "_attn_implementation", "_upcast_and_reordered_attn", "alpha", "and", "as", "attention", "attention_i...
decision_transformer/modeling_decision_transformer.py:DecisionTransformerGPT2MLP
[ -0.00021404866129159927, 0.039893656969070435, 0.03347398713231087, 0.007623356766998768, -0.0009600843768566847, 0.041040025651454926, 0.026251859962940216, -0.01404302567243576, 0.000909930735360831, -0.013183248229324818, 0.02957633137702942, -0.019029732793569565, -0.00006358767859637737...
[ "ACT2FN", "Conv1D", "Dropout", "ModelGPT2MLP", "Module", "__init__", "act", "activation_function", "c_fc", "c_proj", "class", "config", "def", "dropout", "embed_dim", "forward", "hidden_size", "hidden_states", "intermediate_size", "nn", "resid_pdrop", "return", "self", ...
decision_transformer/modeling_decision_transformer.py:DecisionTransformerGPT2Block
[ 0, 0.009848969988524914, 0.020705221220850945, -0.01510921586304903, -0.00009618134936317801, 0.0242866650223732, 0.02271978370845318, 0, -0.0009023559396155179, 0.0035254836548119783, 0.0016368316719308496, 0.020593300461769104, -0.0005735905724577606, 0.011919491924345493, -0.019026419...
[ "False", "GradientCheckpointingLayer", "If", "LayerNorm", "ModelGPT2Attention", "ModelGPT2Block", "ModelGPT2MLP", "None", "True", "ValueError", "__init__", "add_cross_attention", "are", "attention", "attention_mask", "attn", "attn_output", "be", "by", "cache_position", "class...
decision_transformer/modeling_decision_transformer.py:DecisionTransformerGPT2PreTrainedModel
[ -0.00019342546875122935, 0.059870459139347076, 0.004473403561860323, -0.001132418168708682, -0.000682264391798526, 0.0056550572626292706, 0.0333113819360733, -0.002968201646581292, -0.0050360956229269505, 0.001132418168708682, 0.020369460806250572, 0.010184730403125286, -0.002110095927491784...
[ "False", "ModelConfig", "ModelGPT2Attention", "ModelGPT2PreTrainedModel", "PreTrainedModel", "True", "_can_compile_fullgraph", "_init_weights", "and", "base_model_prefix", "bias", "bool", "c_proj", "class", "config", "copy_", "def", "dtype", "elif", "for", "if", "in", "in...
decision_transformer/modeling_decision_transformer.py:DecisionTransformerGPT2Model
[ 0.000011165547221025918, 0.04015675187110901, 0.007975577376782894, -0.009035269729793072, -0.0001908491540234536, 0.015504968352615833, 0.04328005760908127, 0, 0.00649758568033576, -0.0044060880318284035, 0.014501050114631653, 0.010150735266506672, -0.00162439642008394, 0.0245402380824089...
[ "BaseModelOutputWithPastAndCrossAttentions", "Dropout", "DynamicCache", "Embedding", "EncoderDecoderCache", "False", "LayerNorm", "ModelGPT2Block", "ModelGPT2Model", "ModelGPT2PreTrainedModel", "ModuleList", "None", "Setting", "True", "ValueError", "_", "__init__", "add_cross_atten...
decision_transformer/modeling_decision_transformer.py:DecisionTransformerOutput
[ -0.00023333181161433458, 0.032374341040849686, 0.01037346851080656, 0.023482797667384148, -0.0010900691850110888, 0.03898600488901138, 0.053349267691373825, -0.022114867344498634, 0.01379329338669777, -0.0005770954303443432, 0.0220008734613657, 0.017099125310778618, -0.0036193146370351315, ...
[ "ModelOutput", "None", "action_preds", "attentions", "class", "hidden_states", "last_hidden_state", "r", "return_preds", "state_preds" ]
decision_transformer/modeling_decision_transformer.py:DecisionTransformerPreTrainedModel
[ -0.00035274893161840737, 0.028510840609669685, 0.011287965811789036, 0.03118736855685711, -0.0017019226215779781, 0.027463503181934357, 0.01291715633124113, -0.002167405793443322, -0.006633134093135595, 0.017339246347546577, 0.03491123393177986, -0.01920117810368538, -0.002371054608374834, ...
[ "False", "Model", "ModelConfig", "ModelPreTrainedModel", "PreTrainedModel", "base_model_prefix", "class", "config", "main_input_name", "states", "supports_gradient_checkpointing" ]
decision_transformer/modeling_decision_transformer.py:DecisionTransformerModel
[ -0.00016735802637413144, 0.04306914657354355, -0.005101777147501707, 0.01939239166676998, -0.0005707789678126574, 0.05186337232589722, 0.03201999515295029, -0.013134962879121304, 0.0047071645967662334, 0.006623854395002127, 0.04171619191765785, -0.007497639860957861, -0.001507983892224729, ...
[ "Embedding", "LayerNorm", "Linear", "ModelGPT2Model", "ModelModel", "ModelOutput", "ModelPreTrainedModel", "None", "Sequential", "Tanh", "__init__", "act_dim", "action_embeddings", "action_preds", "action_tanh", "actions", "attention_mask", "attentions", "auto_docstring", "batc...
qwen2_vl/modeling_qwen2_vl.py:Qwen2VLModelOutputWithPast
[ -0.0002648343506734818, 0.008071142248809338, 0.014355102553963661, 0.0052174171432852745, -0.0014196563279256225, 0.042200542986392975, 0.04911866411566734, -0.03851087763905525, 0.016949398443102837, 0.0025078190956264734, 0.008071142248809338, 0.03182335942983627, -0.001830419758334756, ...
[ "ModelModelOutputWithPast", "ModelOutput", "None", "attentions", "class", "hidden_states", "last_hidden_state", "past_key_values", "r", "rope_deltas" ]
qwen2_vl/modeling_qwen2_vl.py:Qwen2VLCausalLMOutputWithPast
[ -0.0002988479973282665, 0.017503436654806137, 0.020285440608859062, -0.0002644352207425982, -0.0015503872418776155, 0.049844224005937576, 0.049612391740083694, -0.02874736674129963, 0.016923854127526283, 0.0007607040461152792, 0.017155686393380165, 0.02156052552163601, -0.0010577408829703927...
[ "ModelCausalLMOutputWithPast", "ModelOutput", "None", "attentions", "class", "hidden_states", "logits", "loss", "past_key_values", "r", "rope_deltas" ]
qwen2_vl/modeling_qwen2_vl.py:Qwen2VLRotaryEmbedding
[ -0.000279157335171476, 0.05025552585721016, 0.001159853651188314, -0.006714184302836657, -0.001397587708197534, 0.0389595590531826, 0.03919008746743202, 0.007953282445669174, -0.005129290744662285, 0.021439284086227417, 0.0008824974065646529, -0.000720406009349972, -0.0017577906837686896, ...
[ "False", "ModelRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "arange", "attention_factor", "attention_scaling", "base", "cat", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cpu", "def", "default", "device", ...
qwen2_vl/modeling_qwen2_vl.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
qwen2_vl/modeling_qwen2_vl.py:apply_multimodal_rotary_pos_emb
[ -0.0001921519433381036, 0.016550207510590553, 0.023101331666111946, -0.009366958402097225, -0.0006752255139872432, 0.039766471832990646, 0.031951095908880234, -0.022296808660030365, 0.01022894773632288, 0.04620266333222389, 0.019768305122852325, 0.003376127453520894, -0.00028733001090586185,...
[ "Model_multimodal_rotary_pos_emb", "cat", "cos", "def", "dim", "enumerate", "for", "i", "in", "k", "k_embed", "m", "mrope_section", "q", "q_embed", "return", "rotate_half", "sin", "split", "torch", "unsqueeze", "unsqueeze_dim" ]
qwen2_vl/modeling_qwen2_vl.py:apply_rotary_pos_emb_vision
[ -0.00006463831959990785, 0.03513491153717041, 0.018587501719594002, -0.0022809358779340982, -0.0004639792023226619, 0.01700076460838318, 0.060749396681785583, -0.009180412627756596, 0.016887426376342773, 0.04692211002111435, 0.009010405279695988, 0.01983422413468361, 0.0013458938337862492, ...
[ "Model_rotary_pos_emb_vision", "cos", "def", "dtype", "float", "k", "k_embed", "orig_q_dtype", "q", "q_embed", "return", "rotate_half", "sin", "unsqueeze" ]
qwen2_vl/modeling_qwen2_vl.py:VisionRotaryEmbedding
[ -0.0002629260125104338, 0.03526810184121132, 0.018440837040543556, 0.0027229050174355507, -0.0012245868565514684, 0.04287494719028473, 0.06546497344970703, -0.007894983515143394, 0.00463902298361063, 0.02431885525584221, 0.008183121681213379, -0.01354249007999897, -0.002290697768330574, 0....
[ "False", "ModelRotaryEmbedding", "Module", "Tensor", "__init__", "arange", "class", "def", "device", "dim", "dtype", "float", "forward", "freqs", "inv_freq", "nn", "outer", "persistent", "register_buffer", "return", "self", "seq", "seqlen", "super", "theta", "torch"...
qwen2_vl/modeling_qwen2_vl.py:PatchEmbed
[ -0.00004448119943845086, 0.00798802636563778, 0.023907428607344627, 0.008667859248816967, 0, 0.04464230686426163, 0.026286840438842773, -0.040789924561977386, 0.018582075834274292, 0.0009276875061914325, 0.014049862511456013, -0.00643007829785347, -0.002110311994329095, 0.00827128998935222...
[ "Conv3d", "ModelEmbed", "Model_size", "Module", "__init__", "class", "def", "dtype", "embed_dim", "forward", "hidden_states", "in_channels", "kernel_size", "nn", "proj", "return", "self", "stride", "super", "target_dtype", "temporal_Model_size", "view", "weight" ]
qwen2_vl/modeling_qwen2_vl.py:PatchMerger
[ -0.00022951771097723395, 0.012696421705186367, 0.014176721684634686, 0.016738779842853546, -0.0005906967562623322, 0.04440900683403015, 0.008995670825242996, -0.03689363971352577, 0.007173763122409582, 0.008654063567519188, 0.017308127135038376, 0.008654063567519188, -0.0006013719830662012, ...
[ "GELU", "LayerNorm", "Linear", "ModelMerger", "Module", "Sequential", "__init__", "class", "context_dim", "def", "dim", "eps", "forward", "hidden_size", "ln_q", "mlp", "nn", "return", "self", "spatial_merge_size", "super", "view", "x" ]
qwen2_vl/modeling_qwen2_vl.py:VisionMlp
[ -0.00015563804481644183, 0.017091887071728706, 0.04233355075120926, 0.02286464348435402, -0.00037848344072699547, 0.049125026911497116, 0.03169356659054756, -0.010922960937023163, -0.0006720733945257962, -0.010357004590332508, 0.023317409679293633, -0.035542070865631104, -0.00097627501236274...
[ "ACT2FN", "Linear", "ModelMlp", "Module", "__init__", "act", "class", "def", "dim", "fc1", "fc2", "forward", "hidden_act", "hidden_dim", "nn", "return", "self", "super", "x" ]
qwen2_vl/modeling_qwen2_vl.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
qwen2_vl/modeling_qwen2_vl.py:eager_attention_forward
[ 0, 0.021594731137156487, 0.01775064319372177, -0.018768195062875748, -0.00008832923776935786, 0.039797618985176086, 0.05359111353754997, -0.035049039870500565, 0.02069023996591568, 0.011645326390862465, 0.029395969584584236, 0.022499222308397293, 0.002741739386692643, -0.014584923163056374...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "causal_mask", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "key_states", "kwargs", "matmul", "module", "nn", "not", "num_key_value_groups", ...
qwen2_vl/modeling_qwen2_vl.py:VisionAttention
[ -0.0001046133620548062, 0.028142767027020454, 0.031320177018642426, 0.007603086065500975, -0.0005000163801014423, 0.034951500594615936, 0.02791580930352211, -0.017589230090379715, 0.006099490914493799, 0.009986143559217453, 0.024057526141405106, 0.023376652970910072, 0.0010496798204258084, ...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "ModelAttention", "Module", "None", "Tensor", "_", "__init__", "_attn_implementation", "apply_rotary_pos_emb_Model", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_outputs", "cat", "class", "config"...
qwen2_vl/modeling_qwen2_vl.py:Qwen2VLVisionBlock
[ -0.00020175975805614144, 0.017103563994169235, 0.015970878303050995, 0.0198220107704401, -0.0004052893491461873, 0.043948229402303696, 0.024805830791592598, -0.021860847249627113, 0.00430420832708478, 0.021067965775728226, 0.02288026362657547, 0.016537221148610115, 0.002038835547864437, -0...
[ "GradientCheckpointingLayer", "LayerNorm", "ModelVisionBlock", "None", "Tensor", "VisionAttention", "VisionMlp", "__init__", "attn", "attn_implementation", "class", "config", "cu_seqlens", "def", "dim", "embed_dim", "eps", "forward", "hidden_act", "hidden_dim", "hidden_states...
qwen2_vl/modeling_qwen2_vl.py:Qwen2MLP
[ -0.0002434849739074707, 0.024471141397953033, 0.0228551235049963, 0.026548879221081734, -0.0009450823999941349, 0.060485273599624634, 0.036244992166757584, -0.005078915972262621, 0, -0.005021201446652412, 0.028164898976683617, -0.04640282690525055, -0.0015583038330078125, 0.009176678024232...
[ "ACT2FN", "Linear", "ModelMLP", "Module", "__init__", "act_fn", "class", "config", "def", "down_proj", "forward", "gate_proj", "hidden_act", "hidden_size", "intermediate_size", "nn", "return", "self", "super", "up_proj", "x" ]
qwen2_vl/modeling_qwen2_vl.py:Qwen2VLAttention
[ -0.0000965137078310363, 0.027866460382938385, 0.02865619771182537, -0.005161500070244074, -0.0005112141370773315, 0.031363870948553085, 0.03452282026410103, -0.013312721624970436, 0.0012057602871209383, 0.017148591578006744, 0.015681935474276543, 0.022338295355439186, -0.0011211455566808581,...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Instantiating", "Linear", "ModelAttention", "Module", "None", "Please", "Tensor", "True", "ValueError", "_", "__class__", "__init__", "__name__", "_attn_implementation", "a", "and", "apply_multimodal_rotary_pos_emb", "attention_dropout", "...
qwen2_vl/modeling_qwen2_vl.py:Qwen2VLDecoderLayer
[ -0.00011485910363262519, 0.037546556442976, 0.013514498248696327, -0.000379918550606817, -0.0007845759973861277, 0.037320371717214584, 0.043879710137844086, -0.03980839625000954, 0.007350982632488012, -0.010743743740022182, 0.0016681075794622302, 0.02465406432747841, -0.0018094725674018264, ...
[ "Attention", "False", "GradientCheckpointingLayer", "ModelAttention", "ModelDecoderLayer", "ModelMLP", "ModelRMSNorm", "None", "Sliding", "Tensor", "Window", "__init__", "_attn_implementation", "and", "attention_mask", "attention_type", "be", "but", "cache_position", "class", ...
qwen2_vl/modeling_qwen2_vl.py:Qwen2VLPreTrainedModel
[ -0.0003677352797240019, 0.04730313643813133, 0.0021408747415989637, 0.01433075312525034, -0.001667552045546472, 0.03378795459866524, 0.02971009723842144, -0.025515731424093246, -0.0030875199008733034, 0.022952506318688393, 0.007514907047152519, -0.008213968016207218, -0.004835172556340694, ...
[ "ModelConfig", "ModelDecoderLayer", "ModelPreTrainedModel", "ModelVisionBlock", "PreTrainedModel", "True", "VisionRotaryEmbedding", "_can_compile_fullgraph", "_init_weights", "_no_split_modules", "_skip_keys_device_placement", "_supports_attention_backend", "_supports_flash_attn", "_suppor...
qwen2_vl/modeling_qwen2_vl.py:Qwen2VisionTransformerPretrainedModel
[ -0.0001913632731884718, 0.044473882764577866, 0.006151844281703234, 0.007026647683233023, -0.0009241875959560275, 0.03205731883645058, 0.03273458406329155, -0.05169806629419327, -0.0067726727575063705, 0.016818804666399956, 0.010723398067057133, 0.011457104235887527, -0.0034568849951028824, ...
[ "BaseModelOutputWithPooling", "F", "False", "ModelPreTrainedModel", "ModelVisionBlock", "ModelVisionConfig", "ModelVisionTransformerPretrainedModel", "ModuleList", "PatchEmbed", "PatchMerger", "VisionAttention", "VisionRotaryEmbedding", "_", "__init__", "_can_record_outputs", "_input_e...
qwen2_vl/modeling_qwen2_vl.py:Qwen2VLTextModel
[ 0.000010847413250303362, 0.05112914368510246, 0.015070817433297634, -0.008595948107540607, -0.00027734492323361337, 0.02947182208299637, 0.03773286193609238, -0.020540965721011162, 0.010437936522066593, 0.0006174849113449454, 0.02322022244334221, 0.01440100371837616, -0.0030281180515885353, ...
[ "BaseModelOutputWithPast", "DynamicCache", "Embedding", "False", "ModelDecoderLayer", "ModelPreTrainedModel", "ModelRMSNorm", "ModelRotaryEmbedding", "ModelTextConfig", "ModelTextModel", "ModuleList", "None", "Setting", "True", "ValueError", "You", "__init__", "_attn_implementation...
qwen2_vl/modeling_qwen2_vl.py:Qwen2VLModel
[ 0, 0.03949320316314697, -0.012606588192284107, -0.0016943588852882385, 0, 0.05377323180437088, 0.03079131245613098, -0.04105508327484131, 0.005550245754420757, 0.01807316206395626, 0.033245690166950226, 0.019077226519584656, 0.0006449720240198076, 0.015172530896961689, -0.022647233679890...
[ "False", "Image", "ModelModel", "ModelModelOutputWithPast", "ModelPreTrainedModel", "ModelTextModel", "ModelVisionTransformerPretrainedModel", "None", "True", "Video", "_", "__init__", "_checkpoint_conversion_mapping", "_from_config", "accepts_loss_kwargs", "all", "and", "append", ...
qwen2_vl/modeling_qwen2_vl.py:Qwen2VLForConditionalGeneration
[ -0.0002857950166799128, 0.03408238664269447, -0.011644816026091576, 0.01919974572956562, -0.0010721751023083925, 0.040217217057943344, 0.029538068920373917, -0.05294130742549896, -0.004771534353494644, 0.01772284135222435, 0.03771784156560898, 0.004231896251440048, 0.0012354865903034806, -...
[ "False", "GenerationMixin", "Linear", "ModelCausalLMOutputWithPast", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "Tensor", "True", "__init__", "_checkpoint_conversion_mapping", "_expand_dict_for_generation", "_expand_dict_for_generation_visual", "_expan...
xlnet/modeling_xlnet.py:XLNetRelativeAttention
[ -0.0002689343236852437, 0.03396153077483177, 0.028070244938135147, -0.019175559282302856, -0.0011984724551439285, 0.0399683341383934, 0.03142019361257553, -0.009587779641151428, 0.005631376523524523, 0.02772369794547558, 0.009183474816381931, 0.009934325702488422, 0.0010396386496722698, 0....
[ "Dropout", "False", "FloatTensor", "LayerNorm", "Model", "Module", "None", "Parameter", "The", "True", "ValueError", "__init__", "a", "ac", "arange", "attention", "attn_mask", "attn_mask_g", "attn_mask_h", "attn_out", "attn_prob", "attn_prob_g", "attn_prob_h", "attn_sco...
xlnet/modeling_xlnet.py:XLNetFeedForward
[ -0.00023995348601602018, 0.06051125377416611, 0.020285023376345634, 0.02956799976527691, -0.0008881860412657261, 0.04217451065778732, 0.03208930045366287, -0.010027906857430935, 0, 0.007678511552512646, 0.025098418816924095, -0.019941208884119987, 0.003667348762974143, 0, -0.006790325511...
[ "ACT2FN", "Dropout", "LayerNorm", "Linear", "Model", "Module", "__init__", "activation_function", "class", "config", "d_inner", "d_model", "def", "dropout", "else", "eps", "ff_activation", "forward", "if", "inp", "isinstance", "layer_1", "layer_2", "layer_norm", "laye...
xlnet/modeling_xlnet.py:XLNetLayer
[ -0.0002583168097771704, 0.03903454169631004, 0.0339830107986927, -0.0006852570804767311, -0.0008072400232777, 0.030309172347187996, 0.024224376305937767, -0.02387995459139347, 0.004247876349836588, 0.003028047038242221, 0.007405082229524851, 0.0058551812544465065, 0.006113498006016016, 0.0...
[ "Dropout", "False", "Model", "ModelFeedForward", "ModelRelativeAttention", "Module", "None", "__init__", "apply_chunking_to_forward", "attn_mask_g", "attn_mask_h", "chunk_size_feed_forward", "class", "config", "def", "dropout", "ff", "ff_chunk", "forward", "if", "is", "mems...
xlnet/modeling_xlnet.py:XLNetPoolerStartLogits
[ -0.00019293669902253896, 0.0205988809466362, 0.044611942023038864, 0.02333022467792034, -0.0010313666425645351, 0.01843656785786152, 0.04233582317829132, -0.03163805976510048, 0.007738806307315826, 0.016615672037005424, 0.020371269434690475, -0.010811567306518555, 0.0027597949374467134, 0....
[ "Linear", "Model", "Module", "None", "__init__", "class", "config", "def", "dense", "else", "float16", "forward", "hidden_size", "hidden_states", "if", "is", "nn", "not", "p_mask", "return", "self", "squeeze", "super", "torch", "x" ]
xlnet/modeling_xlnet.py:XLNetPoolerEndLogits
[ -0.0004017357714474201, 0.02606169506907463, 0.029217146337032318, 0.02863280475139618, -0.001665377407334745, 0.04324137791991234, 0.03716421127319336, -0.03926784545183182, 0.0032723203767091036, 0.031320780515670776, 0.030385833233594894, 0.005960298236459494, 0.0028048460371792316, -0....
[ "LayerNorm", "Linear", "Model", "Module", "None", "One", "Tanh", "__init__", "activation", "assert", "be", "cat", "class", "config", "def", "dense_0", "dense_1", "dim", "else", "eps", "expand", "float16", "forward", "gather", "hidden_size", "hidden_states", "hsz",...
xlnet/modeling_xlnet.py:XLNetPoolerAnswerClass
[ -0.00037237905780784786, 0.01581067033112049, 0.02929624170064926, 0.019879592582583427, -0.001634834916330874, 0.04161926358938217, 0.03720157593488693, -0.022669710218906403, -0.001918206224218011, 0.022902220487594604, 0.0339464396238327, 0.020344611257314682, 0.0016203030245378613, -0....
[ "Linear", "Model", "Module", "None", "One", "Tanh", "__init__", "activation", "assert", "be", "cat", "class", "cls_index", "cls_token_state", "config", "def", "dense_0", "dense_1", "dim", "else", "expand", "forward", "gather", "hidden_size", "hidden_states", "hsz", ...
xlnet/modeling_xlnet.py:XLNetSequenceSummary
[ -0.00035416518221609294, 0.03277934715151787, 0.028478512540459633, 0.03812633454799652, -0.0015183696523308754, 0.026502452790737152, 0.01917940564453602, -0.008950389921665192, -0.0032110975589603186, -0.00807859841734171, 0.03510412573814392, 0.012088838033378124, 0.0032256273552775383, ...
[ "Dropout", "Identity", "Linear", "Model", "Module", "None", "NotImplementedError", "__init__", "activation", "activation_string", "and", "attn", "class", "cls_index", "config", "def", "dim", "dtype", "elif", "else", "expand", "first", "first_dropout", "forward", "full...
xlnet/modeling_xlnet.py:XLNetPreTrainedModel
[ -0.00010629191092448309, 0.05035678297281265, 0.012081111781299114, 0.001566592720337212, -0.00040399745921604335, 0.04087254777550697, 0.012758557684719563, -0.003090845188125968, 0.0026109879836440086, 0.00050455576274544, 0.0031049586832523346, 0.011234304867684841, 0.0005998215638101101,...
[ "Model", "ModelConfig", "ModelModel", "ModelRelativeAttention", "PreTrainedModel", "_init_weights", "base_model_prefix", "class", "config", "def", "elif", "for", "if", "in", "init", "initializer_range", "isinstance", "k", "mask_emb", "mean", "module", "no_grad", "normal_"...
xlnet/modeling_xlnet.py:XLNetModelOutput
[ -0.00018809425819199532, 0.019031556323170662, 0.018114373087882996, 0.013929723761975765, -0.001096320804208517, 0.05136227235198021, 0.05319663882255554, -0.019375499337911606, 0.023846769705414772, -0.005273804999887943, 0.0005409949226304889, 0.01616535894572735, -0.0010246658930554986, ...
[ "FloatTensor", "Model", "ModelOutput", "None", "attentions", "class", "hidden_states", "last_hidden_state", "mems", "r", "torch" ]
xlnet/modeling_xlnet.py:XLNetLMHeadModelOutput
[ -0.0001714007230475545, 0.016568735241889954, 0.007198830135166645, 0.004884920548647642, -0.0010998212965205312, 0.06353253126144409, 0.05279142037034035, -0.01051257736980915, 0.021367955952882767, -0.01542606484144926, 0.0033280306961387396, 0.011083913035690784, -0.0016211650799959898, ...
[ "Model", "ModelOutput", "None", "attentions", "class", "hidden_states", "logits", "loss", "mems", "r" ]
xlnet/modeling_xlnet.py:XLNetForSequenceClassificationOutput
[ -0.0001714007230475545, 0.016568735241889954, 0.007198830135166645, 0.004884920548647642, -0.0010998212965205312, 0.06353253126144409, 0.05279142037034035, -0.01051257736980915, 0.021367955952882767, -0.01542606484144926, 0.0033280306961387396, 0.011083913035690784, -0.0016211650799959898, ...
[ "Model", "ModelOutput", "None", "attentions", "class", "hidden_states", "logits", "loss", "mems", "r" ]
xlnet/modeling_xlnet.py:XLNetForTokenClassificationOutput
[ -0.0001714007230475545, 0.016568735241889954, 0.007198830135166645, 0.004884920548647642, -0.0010998212965205312, 0.06353253126144409, 0.05279142037034035, -0.01051257736980915, 0.021367955952882767, -0.01542606484144926, 0.0033280306961387396, 0.011083913035690784, -0.0016211650799959898, ...
[ "Model", "ModelOutput", "None", "attentions", "class", "hidden_states", "logits", "loss", "mems", "r" ]
xlnet/modeling_xlnet.py:XLNetForMultipleChoiceOutput
[ -0.0001714007230475545, 0.016568735241889954, 0.007198830135166645, 0.004884920548647642, -0.0010998212965205312, 0.06353253126144409, 0.05279142037034035, -0.01051257736980915, 0.021367955952882767, -0.01542606484144926, 0.0033280306961387396, 0.011083913035690784, -0.0016211650799959898, ...
[ "Model", "ModelOutput", "None", "attentions", "class", "hidden_states", "logits", "loss", "mems", "r" ]
xlnet/modeling_xlnet.py:XLNetForQuestionAnsweringSimpleOutput
[ -0.00021028745686635375, 0.0253132414072752, 0.016264617443084717, 0.009392243809998035, -0.001138236839324236, 0.06643294543027878, 0.05497898533940315, -0.006528754718601704, 0.020044421777129173, -0.003006663406267762, 0.006757833529263735, 0.02050258032977581, 0.000015659705240977928, ...
[ "Model", "ModelOutput", "None", "attentions", "class", "end_logits", "hidden_states", "loss", "mems", "r", "start_logits" ]
xlnet/modeling_xlnet.py:XLNetForQuestionAnsweringOutput
[ -0.00016630563186481595, 0.019579598680138588, 0.027092698961496353, 0.01013130322098732, -0.001038743183016777, 0.06784558296203613, 0.05760044604539871, -0.006517046596854925, 0.016733726486563683, -0.007797688711434603, 0.011668073944747448, 0.027548039332032204, -0.0007470413693226874, ...
[ "Model", "ModelOutput", "None", "attentions", "class", "cls_logits", "end_top_index", "end_top_log_probs", "hidden_states", "loss", "mems", "r", "start_top_index", "start_top_log_probs" ]
xlnet/modeling_xlnet.py:XLNetModel
[ -0.00029076257487758994, 0.03308231756091118, 0.015277598984539509, -0.004709635861217976, -0.0011415123008191586, 0.05146138370037079, 0.017000636085867882, -0.009878748096525669, 0.004910656716674566, -0.007811103016138077, 0.0033886402379721403, 0.009419271722435951, 0.0011271536350250244...
[ "BERT", "Dropout", "Embedding", "FloatTensor", "Model", "ModelLayer", "ModelOutput", "ModelPreTrainedModel", "ModuleList", "None", "Parameter", "Please", "Unsupported", "ValueError", "You", "_", "__init__", "added", "and", "arange", "assert", "att_stream", "attention", ...
xlnet/modeling_xlnet.py:XLNetLMHeadModel
[ -0.0002483236021362245, 0.03700730949640274, 0.02281739003956318, -0.002454856177791953, -0.0010642439592629671, 0.042683277279138565, 0.02293091081082821, -0.02554185502231121, -0.0009578195749782026, -0.0035332900006324053, 0.0019440189935266972, 0.026109451428055763, 0.0004860047483816743...
[ "CrossEntropyLoss", "False", "GenerationMixin", "Linear", "Model", "ModelModel", "ModelOutput", "ModelPreTrainedModel", "None", "__init__", "_reorder_cache", "_tied_weights_keys", "attention_mask", "attentions", "attn_type", "auto_docstring", "beam_idx", "cat", "class", "config...
xlnet/modeling_xlnet.py:XLNetForSequenceClassification
[ -0.00039184876368381083, 0.03651958331465721, 0.016641834750771523, 0.015717288479208946, -0.0014301578048616648, 0.04715186730027199, 0.015370584093034267, 0.008840975351631641, -0.007194126956164837, -0.001271251356229186, 0.02854537032544613, 0.006211796309798956, 0.00034851065720431507, ...
[ "BCEWithLogitsLoss", "CrossEntropyLoss", "Linear", "MSELoss", "Model", "ModelModel", "ModelOutput", "ModelPreTrainedModel", "ModelSequenceSummary", "None", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "class", "config", "d_model", "def", "dtype", "elif...
xlnet/modeling_xlnet.py:XLNetForTokenClassification
[ -0.0002555835817474872, 0.03975744545459747, 0.019083572551608086, 0.009257804602384567, -0.000951338850427419, 0.03748558834195137, 0.025899134576320648, 0.009768972173333168, 0.002555835759267211, 0.0067871636711061, 0.025558356195688248, 0.008860230445861816, 0.001128827454522252, 0.007...
[ "CrossEntropyLoss", "Linear", "Model", "ModelModel", "ModelOutput", "ModelPreTrainedModel", "None", "__init__", "attention_mask", "attentions", "auto_docstring", "class", "classifier", "config", "def", "else", "forward", "hidden_size", "hidden_states", "if", "input_ids", "i...
xlnet/modeling_xlnet.py:XLNetForMultipleChoice
[ -0.00026316341245546937, 0.0464305616915226, 0.02014267072081566, 0.025377489626407623, -0.001081103808246553, 0.04438215494155884, 0.034367721527814865, 0.005035667680203915, -0.00044275468098931015, 0.00842122919857502, 0.024922287091612816, 0.005889170803129673, -0.00015469742356799543, ...
[ "CrossEntropyLoss", "Linear", "Model", "ModelModel", "ModelOutput", "ModelPreTrainedModel", "ModelSequenceSummary", "None", "__init__", "attention_mask", "attentions", "auto_docstring", "class", "config", "d_model", "def", "else", "flat_attention_mask", "flat_input_ids", "flat_...
xlnet/modeling_xlnet.py:XLNetForQuestionAnsweringSimple
[ -0.0003025654295925051, 0.03531116619706154, 0.011219837702810764, 0.01981981471180916, -0.0009753285557962954, 0.057181306183338165, 0.019478093832731247, 0.015491350553929806, 0.0019506571115925908, 0.024831723421812057, 0.016516512259840965, 0.019136372953653336, 0.0014594332315027714, ...
[ "CrossEntropyLoss", "Linear", "Model", "ModelModel", "ModelOutput", "ModelPreTrainedModel", "None", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "clamp", "class", "config", "contiguous", "def", "dim", "else", "end_logits", "end_loss", "end_positions"...
xlnet/modeling_xlnet.py:XLNetForQuestionAnswering
[ -0.0002239142922917381, 0.03912457823753357, 0.023315519094467163, 0.027296219021081924, -0.0008601152221672237, 0.0705152302980423, 0.03980698436498642, 0.01319317240267992, 0.00449250265955925, 0.006511285435408354, 0.022291911765933037, 0.015695326030254364, 0.0031561253126710653, 0.001...
[ "BCEWithLogitsLoss", "CrossEntropyLoss", "Model", "ModelModel", "ModelOutput", "ModelPoolerAnswerClass", "ModelPoolerEndLogits", "ModelPoolerStartLogits", "ModelPreTrainedModel", "None", "__init__", "and", "answer_class", "attention_mask", "attentions", "auto_docstring", "bh", "bl"...
vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackbonePatchEmbeddings
[ -0.0000447696256742347, 0.011910475790500641, 0.02280968427658081, 0.015281365253031254, 0, 0.0021629873663187027, 0.01275319792330265, -0.010281212627887726, 0.008314860053360462, 0.004325974732637405, 0.011685749515891075, 0.007977771572768688, -0.00279502896592021, -0.004775426350533962...
[ "Conv2d", "Input", "Iterable", "ModelPatchEmbeddings", "Module", "ValueError", "__init__", "abc", "class", "collections", "config", "def", "doesn", "else", "embed_dim", "embeddings", "f", "flatten", "forward", "height", "hidden_size", "if", "image", "image_size", "isi...
vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneEmbeddings
[ -0.00010482146899448708, 0.011901403777301311, 0.01235051266849041, 0.012631206773221493, -0.00039647953235544264, 0.01650477573275566, 0.024251915514469147, -0.0038174311630427837, 0.007691001053899527, 0.004013916477560997, 0.013248732313513756, 0.017066163942217827, -0.0004263032169546932...
[ "Dropout", "ModelEmbeddings", "ModelPatchEmbeddings", "Module", "Parameter", "__init__", "class", "config", "def", "dropout", "embeddings", "forward", "hidden_dropout_prob", "hidden_size", "nn", "num_patches", "patch_embeddings", "pixel_values", "position_embeddings", "return",...
vitpose_backbone/modeling_vitpose_backbone.py:eager_attention_forward
[ 0.00002095530362566933, 0.025862814858555794, 0.0269921962171793, -0.01146321278065443, 0.00008955634984886274, 0.03704368323087692, 0.060083046555519104, -0.023942869156599045, 0.021458230912685394, 0.014569009654223919, 0.023152301087975502, 0.026879258453845978, 0.003063444746658206, -0...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "shape", "size", "softmax", "tor...
vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneSelfAttention
[ -0.00010323112655896693, 0.03958452492952347, 0.044982410967350006, 0.00047090963926166296, -0.0005376804037950933, 0.012370163574814796, 0.028788743540644646, -0.020242085680365562, 0.0002301834465470165, 0.021928926929831505, 0.017430685460567474, 0.007478326093405485, -0.00055525166681036...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "ModelSelfAttention", "Module", "None", "The", "ValueError", "__init__", "_attn_implementation", "a", "all_head_size", "and", "attention", "attention_head_size", "attention_interface", "attention_probs", "attention_probs_dropout_prob", ...
vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneSelfOutput
[ -0.000045805918489350006, 0.04578312486410141, 0.04645640775561333, 0.005386250093579292, -0.000490934238769114, 0.04219229146838188, 0.028277814388275146, -0.018739663064479828, 0.0011501888511702418, 0.010940820910036564, 0.010323646478354931, 0.0002261804329464212, 0.003506673267111182, ...
[ "Dropout", "Linear", "ModelSelfOutput", "Module", "__init__", "class", "config", "def", "dense", "dropout", "forward", "hidden_dropout_prob", "hidden_size", "hidden_states", "input_tensor", "nn", "return", "self", "super" ]
vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneAttention
[ 0.000160034658620134, 0.038267627358436584, 0.04974791780114174, -0.0009707597200758755, 0.0004537246422842145, 0.028475617989897728, 0.035791490226984024, -0.014631740748882294, 0.01390015333890915, -0.004361384082585573, 0.0227354746311903, 0.02566182240843773, 0.004699039738625288, -0.0...
[ "ModelAttention", "ModelSelfAttention", "ModelSelfOutput", "Module", "_", "__init__", "attention", "class", "config", "def", "forward", "hidden_states", "nn", "output", "return", "self", "self_attn_output", "super" ]
vitpose_backbone/modeling_vitpose_backbone.py:VitPoseNaiveMoe
[ -0.0003564964863471687, 0.023394854739308357, 0.007470139302313328, 0.02988055720925331, -0.001150922616943717, 0.042157065123319626, 0.044705022126436234, -0.03150198236107826, -0.005182771012187004, -0.007470139302313328, 0.013492577709257603, -0.012913497164845467, -0.0011798766208812594,...
[ "Linear", "ModelPoseNaiveMoe", "ModuleList", "_", "__init__", "append", "class", "config", "current_hidden_state", "def", "expert_hidden_state", "for", "forward", "hidden_features", "hidden_size", "hidden_state", "i", "in", "indices", "int", "mlp_ratio", "nn", "num_expert...
vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneMoeMLP
[ -0.00043258527875877917, 0.03628562018275261, 0.016375653445720673, 0.01873186230659485, -0.0015609884867444634, 0.06314640492200851, 0.04005555436015129, -0.026860784739255905, -0.0036668505053967237, -0.008305637165904045, 0.02155931293964386, -0.016257843002676964, -0.000192362378584221, ...
[ "ACT2FN", "Linear", "ModelMoeMLP", "ModelPoseNaiveMoe", "Module", "__init__", "act", "cat", "class", "config", "def", "dim", "expert_hidden_state", "experts", "fc1", "fc2", "forward", "hidden_act", "hidden_features", "hidden_size", "hidden_state", "in_features", "indices"...
vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneMLP
[ -0.00023419586068484932, 0.024828337132930756, 0.04210519790649414, 0.019565187394618988, -0.0007723100134171546, 0.04919900745153427, 0.030434735119342804, -0.030663566663861275, 0.0038329458329826593, -0.00989700946956873, 0.03272306174039841, -0.03226539492607117, 0.00043263661791570485, ...
[ "ACT2FN", "Linear", "ModelMLP", "Module", "__init__", "activation", "class", "config", "def", "fc1", "fc2", "forward", "hidden_act", "hidden_features", "hidden_size", "hidden_state", "in_features", "int", "mlp_ratio", "nn", "out_features", "return", "self", "super" ]
vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneLayer
[ -0.0001844393409555778, 0.015588212758302689, 0.006127974949777126, 0.022365696728229523, -0.00039358827052637935, 0.06099735572934151, 0.019428787752985954, -0.0083024175837636, -0.0004624220891855657, 0.002696873852983117, 0.010166225954890251, -0.0020614846143871546, 0.0008754250011406839...
[ "GradientCheckpointingLayer", "LayerNorm", "ModelAttention", "ModelLayer", "ModelMLP", "ModelMoeMLP", "None", "Please", "ValueError", "__init__", "and", "attention", "attention_output", "be", "class", "config", "dataset_index", "def", "else", "eps", "experts", "f", "forwa...
vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneEncoder
[ -0.00017254968406632543, 0.013704869896173477, 0.015290557406842709, 0.03579122945666313, -0.00029554663342423737, 0.04757062345743179, 0.012119182385504246, -0.032393328845500946, 0.006937382277101278, 0.014950767159461975, 0.01302528940141201, -0.010250336490571499, -0.0003716454957611859,...
[ "BaseModelOutput", "False", "ModelEncoder", "ModelLayer", "Module", "ModuleList", "None", "_", "__init__", "all_hidden_states", "class", "config", "dataset_index", "def", "else", "enumerate", "for", "forward", "gradient_checkpointing", "hidden_states", "i", "if", "in", ...
vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackbonePreTrainedModel
[ -0.00016535783652216196, 0.040530260652303696, 0.002378338249400258, 0.028146015480160713, -0.0007071686559356749, 0.026006918400526047, 0.013678963296115398, -0.01007627323269844, 0.0011821326334029436, 0.014748511835932732, 0.005572910886257887, 0.0014002642128616571, -0.001477665849961340...
[ "Conv2d", "LayerNorm", "Linear", "Model", "ModelConfig", "ModelEmbeddings", "ModelLayer", "ModelPreTrainedModel", "ModelSelfAttention", "None", "PreTrainedModel", "True", "_can_record_outputs", "_init_weights", "_no_split_modules", "_supports_flash_attn", "_supports_sdpa", "attenti...
vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackbone
[ -0.00014467901200987399, 0.01341218315064907, 0.02143704704940319, 0.051628489047288895, -0.0003805496671702713, 0.030977092683315277, 0.01863115094602108, -0.020314687862992287, 0.004012431483715773, 0.010269579477608204, 0.010718523524701595, 0.01060628704726696, 0.000345475971698761, -0...
[ "BackboneMixin", "BackboneOutput", "False", "LayerNorm", "Model", "ModelEmbeddings", "ModelEncoder", "ModelPreTrainedModel", "None", "True", "_", "__init__", "_init_backbone", "append", "auto_docstring", "check_model_inputs", "class", "config", "dataset_index", "def", "else",...
mpt/modeling_mpt.py:build_mpt_alibi_tensor
[ -0.0003601185162551701, 0.019920524209737778, 0.05628706514835358, -0.004169412422925234, -0.0012088400544598699, 0.05026458203792572, 0.02582719177007675, -0.02571137621998787, 0.00330078462138772, -0.009439085610210896, 0.02768026478588581, -0.021194512024521828, -0.0000615277822362259, ...
[ "Model_Model_alibi_tensor", "None", "alibi", "alibi_bias_max", "arange", "base", "ceil", "concat", "def", "device", "dim", "dtype", "float", "if", "int32", "int64", "log2", "math", "num_heads", "num_heads_power_of_2", "pow", "r", "return", "sequence_length", "slopes",...
mpt/modeling_mpt.py:MptAttention
[ -0.00015231447468977422, 0.05048740655183792, 0.036963995546102524, -0.019158167764544487, -0.0008029526215977967, 0.014537668786942959, 0.027722995728254318, -0.02073589898645878, 0.0025919873733073473, 0.014143235050141811, 0.01960894837975502, 0.03944328799843788, 0.0018172085983678699, ...
[ "Expecting", "Linear", "ModelAttention", "Module", "None", "ValueError", "Wqkv", "__init__", "attention_mask", "attention_scores", "attn_config", "attn_dropout_p", "attn_output", "attn_pdrop", "attn_weights", "batch_size", "be", "cache_kwargs", "cache_position", "chunk", "cla...
mpt/modeling_mpt.py:MptMLP
[ -0.0002713424910325557, 0.029096541926264763, 0.03818202391266823, 0.012420658022165298, -0.0010206791339442134, 0.04784253612160683, 0.033811792731285095, -0.016215858981013298, -0.004341479856520891, -0.012190645560622215, 0.027371451258659363, -0.02162114530801773, -0.0020413582678884268,...
[ "F", "GELU", "Linear", "ModelMLP", "Module", "__init__", "act", "approximate", "attn_config", "attn_pdrop", "class", "config", "def", "down_proj", "dropout", "forward", "hidden_dropout", "hidden_size", "hidden_states", "intermediate_output", "nn", "none", "output", "p",...
mpt/modeling_mpt.py:MptBlock
[ -0.0000920576203498058, 0.002903559012338519, 0.02559642307460308, -0.014997023157775402, -0.0002942320134025067, 0.04329967498779297, 0.024581586942076683, -0.024130547419190407, 0.004172103945165873, 0.004454003181308508, -0.0016772985691204667, 0.008400588296353817, 0.0017477733781561255,...
[ "Dropout", "False", "GradientCheckpointingLayer", "LayerNorm", "ModelAttention", "ModelBlock", "ModelMLP", "None", "__init__", "attention_mask", "attn", "attn_config", "attn_outputs", "attn_pdrop", "attn_weights", "bias", "cache_position", "class", "config", "def", "dropout_r...
mpt/modeling_mpt.py:MptPreTrainedModel
[ -0.0003524611529428512, 0.02232496254146099, 0.022092411294579506, 0.014069377444684505, -0.001642396324314177, 0.03837103024125099, 0.020929653197526932, -0.00543589610606432, -0.004156861454248428, -0.003531878814101219, 0.005668447818607092, -0.012790343724191189, -0.001976689323782921, ...
[ "ModelBlock", "ModelConfig", "ModelPreTrainedModel", "PreTrainedModel", "True", "_no_split_modules", "base_model_prefix", "class", "config", "supports_gradient_checkpointing", "transformer" ]
mpt/modeling_mpt.py:MptModel
[ -0.0001255102251889184, 0.035500962287187576, 0.009436964988708496, 0.0016079351771622896, -0.00043884693877771497, 0.037747859954833984, 0.02314303256571293, -0.014941860921680927, -0.0020643360912799835, -0.008706723339855671, 0.02831089496612549, -0.004072499927133322, -0.0009338663076050...
[ "BaseModelOutputWithPastAndCrossAttentions", "DynamicCache", "Embedding", "False", "LayerNorm", "ModelBlock", "ModelModel", "ModelPreTrainedModel", "ModuleList", "None", "Setting", "True", "_", "__init__", "_prepare_4d_causal_attention_mask", "alibi", "alibi_bias_max", "all_hidden_...
mpt/modeling_mpt.py:MptForCausalLM
[ -0.00023333256831392646, 0.050682663917541504, 0.017761558294296265, -0.0032242320012301207, -0.0011383801465854049, 0.01979791559278965, 0.03416554629802704, -0.013745409436523914, 0.0020080741960555315, 0.01719590276479721, 0.019232260063290596, -0.0006575736333616078, 0.000562119355890899...
[ "CausalLMOutputWithCrossAttentions", "GenerationMixin", "Linear", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "cache_position", "class", "config", "def", "else", "forward", "...
mpt/modeling_mpt.py:MptForSequenceClassification
[ -0.0003949534730054438, 0.044349685311317444, 0.006922457367181778, 0.019187558442354202, -0.0013428418897092342, 0.016774751245975494, 0.01252361573278904, -0.016200274229049683, -0.005888397339731455, -0.005601158365607262, 0.042970940470695496, 0.01964714005589485, 0.0004129059088882059, ...
[ "BCEWithLogitsLoss", "Cannot", "CrossEntropyLoss", "Linear", "MSELoss", "ModelForSequenceClassification", "ModelModel", "ModelPreTrainedModel", "None", "Results", "SequenceClassifierOutputWithPast", "ValueError", "__class__", "__init__", "__name__", "and", "arange", "argmax", "at...
mpt/modeling_mpt.py:MptForTokenClassification
[ -0.00028813740937039256, 0.037336915731430054, 0.008252824656665325, -0.02959633618593216, -0.0009960305178537965, 0.03187297657132149, 0.0327836349606514, 0.0034861068706959486, -0.0044963667169213295, 0.005435481201857328, 0.056460704654455185, 0.01809929870069027, -0.0003005877952091396, ...
[ "CrossEntropyLoss", "Dropout", "Linear", "ModelForTokenClassification", "ModelModel", "ModelPreTrainedModel", "None", "TokenClassifierOutput", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "batch_size", "class", "classifier", "classifier_dropout", "config", ...
mpt/modeling_mpt.py:MptForQuestionAnswering
[ -0.00019118325144518167, 0.029771726578474045, 0.02199047990143299, 0.0166902095079422, -0.0006378648686222732, 0.05142389237880707, 0.04420650377869606, 0.01815624162554741, 0.007048230618238449, 0.026614118367433548, 0.02052444778382778, 0.017141297459602356, 0.0016281412681564689, 0.002...
[ "CrossEntropyLoss", "Linear", "ModelForQuestionAnswering", "ModelModel", "ModelPreTrainedModel", "None", "QuestionAnsweringModelOutput", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "clamp", "class", "config", "contiguous", "def", "dim", "else", "end_log...
mamba/modeling_mamba.py:MambaCache
[ -0.00018014493980444968, 0.024301905184984207, 0.009551214054226875, -0.01458114292472601, -0.00044153170892968774, 0.006697152741253376, -0.013733401894569397, -0.05222084000706673, 0.006188508588820696, 0.031648993492126465, 0.012998693622648716, -0.0025573517195880413, 0.00116564368363469...
[ "ModelCache", "None", "True", "_", "__init__", "_dtype", "_dynamo", "append", "cache_position", "clamp", "class", "config", "conv_kernel", "conv_kernel_size", "conv_state", "conv_states", "def", "device", "dims", "dtype", "else", "float16", "for", "if", "in", "inter...
mamba/modeling_mamba.py:MambaMixer
[ -0.00015878370322752744, 0.028158824890851974, 0.006330058444291353, -0.0029521349351853132, -0.00022531318245455623, 0.04110280051827431, 0.013000748120248318, -0.03633396700024605, 0.007153249811381102, 0.011978855356574059, 0.01930242031812668, -0.0036617827136069536, 0.001625093515031039...
[ "A", "ACT2FN", "A_log", "B", "C", "Conv1d", "D", "Falling", "False", "ImportError", "Linear", "Model", "ModelMixer", "Model_inner_fn", "Model_ssm", "Modelpy", "Module", "None", "Parameter", "The", "To", "True", "_", "__init__", "act", "activation", "all", "and",...
mamba/modeling_mamba.py:MambaRMSNorm
[ -0.00010958517668768764, 0.04162822663784027, 0.031221171841025352, 0.054750170558691025, -0.0005620659212581813, 0.039139583706855774, 0.021492835134267807, -0.028053805232048035, 0.007465932052582502, 0.043664392083883286, 0.01923043094575405, 0.005882249679416418, 0.0027431645430624485, ...
[ "ModelRMSNorm", "Module", "Parameter", "True", "__init__", "class", "def", "eps", "extra_repr", "f", "float32", "forward", "hidden_size", "hidden_states", "keepdim", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "shape", "super", "to", "torch", "variance...
mamba/modeling_mamba.py:MambaBlock
[ -0.00018491271475795656, -0.0019252155907452106, 0.02095087617635727, 0.004699791315943003, -0.0003945984353777021, 0.05096159130334854, 0.02672652341425419, -0.03623935207724571, 0.004190175328403711, 0.015401724725961685, -0.005350966937839985, -0.007587614469230175, 0.000422910408815369, ...
[ "GradientCheckpointingLayer", "ModelBlock", "ModelMixer", "ModelRMSNorm", "None", "__init__", "attention_mask", "cache_params", "cache_position", "class", "config", "def", "dtype", "eps", "float32", "forward", "hidden_size", "hidden_states", "if", "layer_idx", "layer_norm_eps...
mamba/modeling_mamba.py:MambaPreTrainedModel
[ -0.0003291105676908046, 0.029534095898270607, -0.00658221123740077, 0.004321364685893059, -0.0012949785450473428, 0.029534095898270607, 0.009672989137470722, -0.003977945074439049, -0.010760484263300896, 0.020605184137821198, 0.0009944862686097622, -0.0334261879324913, -0.0026471936143934727...
[ "A", "A_log", "D", "Embedding", "Linear", "ModelBlock", "ModelConfig", "ModelMixer", "ModelPreTrainedModel", "ModelRMSNorm", "None", "PreTrainedModel", "True", "_init_weights", "_is_stateful", "_no_split_modules", "a", "arange", "backbone", "base_model_prefix", "bias", "cla...
mamba/modeling_mamba.py:MambaOutput
[ -0.00017384765669703484, 0.02076900005340576, 0.031952306628227234, 0.02179603837430477, -0.001112625002861023, 0.03309345990419388, 0.05135192349553108, -0.01997019164264202, 0.017916115000844002, -0.016204385086894035, -0.003166701877489686, 0.007988076657056808, -0.0008736958843655884, ...
[ "ModelOutput", "None", "cache_params", "class", "hidden_states", "last_hidden_state", "r" ]
mamba/modeling_mamba.py:MambaCausalLMOutput
[ -0.0001988249714486301, 0.028416814282536507, 0.035378362983465195, 0.004079923965036869, -0.0011483702110126615, 0.03857382759451866, 0.050899192690849304, -0.007817476987838745, 0.01791742816567421, -0.015977324917912483, 0.0023538023233413696, -0.0003940835886169225, -0.002425129758194089...
[ "ModelCausalLMOutput", "ModelOutput", "None", "cache_params", "class", "hidden_states", "logits", "loss", "r" ]
mamba/modeling_mamba.py:MambaModel
[ -0.0000467252975795418, 0.03639044985175133, -0.0022463242057710886, 0.006598577369004488, -0.00010003161878557876, 0.037738244980573654, 0.027405153959989548, 0, 0.003762593027204275, 0.007637502159923315, 0.012186308391392231, -0.012747889384627342, 0.0004282055306248367, 0.0223509259521...
[ "Embedding", "False", "ModelBlock", "ModelCache", "ModelModel", "ModelOutput", "ModelPreTrainedModel", "ModelRMSNorm", "ModuleList", "None", "True", "ValueError", "You", "__init__", "_register_load_state_dict_pre_hook", "a", "all_hidden_states", "and", "arange", "are", "args"...
mamba/modeling_mamba.py:MambaForCausalLM
[ -0.00022307969629764557, 0.030140545219182968, 0.014617031440138817, -0.00611875718459487, -0.0008994006784632802, 0.019602686166763306, 0.012747411616146564, -0.007648446597158909, -0.0009418920380994678, 0.007875067181885242, 0.012917377054691315, -0.013200652785599232, 0.00167840917129069...
[ "Any", "CrossEntropyLoss", "False", "GenerationMixin", "Linear", "ModelCache", "ModelCausalLMOutput", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "Model_outputs", "None", "True", "__init__", "_tied_weights_keys", "_update_model_kwargs_for_generation", "and", "arange",...
internvl/modeling_internvl.py:InternVLVisionRMSNorm
[ -0.00012181837519165128, 0.0503939688205719, 0.020790336653590202, 0.06327493488788605, -0.00037781352875754237, 0.0336713045835495, 0.02259819209575653, -0.03276737779378891, 0.007796376012265682, 0.05536556988954544, 0.02180725522339344, 0.014575833454728127, 0.0037287017330527306, 0.012...
[ "ModelVisionRMSNorm", "Module", "Parameter", "True", "__init__", "class", "def", "eps", "extra_repr", "f", "float32", "forward", "hidden_size", "hidden_states", "keepdim", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "shape", "super", "to", "torch", "tu...
internvl/modeling_internvl.py:eager_attention_forward
[ 0.00006051955642760731, 0.025331631302833557, 0.028271909803152084, -0.014927568845450878, 0.00017228195792995393, 0.039806850254535675, 0.06106732785701752, -0.026349419727921486, 0.022052090615034103, 0.018433285877108574, 0.02861117385327816, 0.02431384287774563, 0.002558607840910554, -...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "causal_mask", "contiguous", "def", "dim", "dropout", "functional", "if", "is", "key", "key_states", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "shap...
internvl/modeling_internvl.py:InternVLVisionAttention
[ -0.00011342836660332978, 0.04862120375037193, 0.029487859457731247, 0.02149687334895134, -0.00019344373140484095, 0.00917274970561266, 0.032414134591817856, -0.02037138305604458, -0.0052335322834551334, 0.05019688978791237, 0.024985896423459053, 0.01857059821486473, -0.002447942504659295, ...
[ "ALL_ATTENTION_FUNCTIONS", "Dropout", "False", "Identity", "Linear", "ModelVisionAttention", "ModelVisionRMSNorm", "Module", "None", "ValueError", "_", "__init__", "_attn_implementation", "and", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_we...
internvl/modeling_internvl.py:InternVLVisionModelOutputWithPooling
[ -0.00007458389882231131, 0.017858019098639488, 0.011736873537302017, 0.04245491325855255, -0.00007984865078469738, 0.040657877922058105, 0.04200565442442894, -0.004296032711863518, 0.01561172679066658, 0.009041323326528072, 0.03391900286078453, 0.017633389681577682, 0.0006317695369943976, ...
[ "BaseModelOutputWithPooling", "ModelVisionModelOutputWithPooling", "class", "r" ]
internvl/modeling_internvl.py:InternVLVisionPatchEmbeddings
[ -0.00002250772013212554, 0.005509011447429657, 0.013884957879781723, 0.016976749524474144, -0.00009617933392291889, -0.0035977219231426716, 0.02259819023311138, -0.01618974842131138, 0.011130452156066895, 0.007982444949448109, 0.013547671027481556, 0.003133953083306551, -0.002796666463837027...
[ "Conv2d", "Make", "ModelVisionPatchEmbeddings", "Module", "ValueError", "__init__", "batch_size", "channel", "class", "config", "configuration", "def", "dimension", "dtype", "embeddings", "flatten", "forward", "height", "hidden_size", "if", "image_size", "in", "kernel_siz...
internvl/modeling_internvl.py:InternVLVisionEmbeddings
[ -0.00023984175641089678, 0.027970878407359123, 0.014781358651816845, 0.01108601875603199, -0.0012294111074879766, 0.019215766340494156, 0.026833850890398026, -0.031381960958242416, 0.0031268259044736624, 0.00943732913583517, 0.028198285028338432, 0.037294503301382065, -0.0005720669869333506,...
[ "Dropout", "False", "Iterable", "ModelVisionEmbeddings", "ModelVisionPatchEmbeddings", "Module", "None", "Parameter", "_", "__init__", "abc", "align_corners", "and", "batch_size", "bicubic", "bool_masked_pos", "cat", "class", "class_pos_embed", "cls_token", "cls_tokens", "c...
internvl/modeling_internvl.py:InternVLVisionMLP
[ -0.0001662031572777778, 0.03708730265498161, 0.032536715269088745, 0.034129418432712555, -0.00047283468302339315, 0.04709859937429428, 0.03162659704685211, -0.03481200709939003, 0.003341838950291276, 0, 0.049373894929885864, -0.020136358216404915, 0.0019482210045680404, -0.0018913387320935...
[ "ACT2FN", "Linear", "ModelVisionMLP", "Module", "__init__", "activation_fn", "class", "config", "def", "fc1", "fc2", "forward", "hidden_act", "hidden_size", "hidden_states", "intermediate_size", "nn", "return", "self", "super" ]
internvl/modeling_internvl.py:InternVLVisionLayer
[ -0.00010581125388853252, 0.0405195914208889, 0.01723761484026909, 0.03313203901052475, 0.0001010016567306593, 0.0405195914208889, 0.02540869265794754, -0.01925240270793438, 0.0019028537208214402, 0.034027501940727234, 0.029102468863129616, 0.01981206424534321, 0.004645201843231916, -0.0154...
[ "Dropout", "GradientCheckpointingLayer", "ModelVisionAttention", "ModelVisionLayer", "ModelVisionMLP", "NORM2FN", "None", "Parameter", "True", "_", "__init__", "attention", "attention_output", "chunk_size_feed_forward", "class", "config", "def", "dropout", "eps", "forward", "...
internvl/modeling_internvl.py:InternVLVisionEncoder
[ -0.00009011043584905565, 0.0219429899007082, 0.011646663770079613, 0.058064527809619904, 0.00009626431710785255, 0.03780945762991905, 0.02160540595650673, -0.0427606962621212, 0.007933234795928001, 0.023405855521559715, 0.027906982228159904, -0.010296326130628586, -0.00014329756959341466, ...
[ "BaseModelOutput", "False", "ModelVisionEncoder", "ModelVisionLayer", "Module", "ModuleList", "__init__", "class", "config", "def", "for", "forward", "gradient_checkpointing", "hidden_states", "i", "in", "last_hidden_state", "layer", "layer_module", "nn", "num_hidden_layers",...