identifier
stringlengths
24
102
embedding
listlengths
2.56k
2.56k
tokens
listlengths
4
448
m2m_100/modeling_m2m_100.py:M2M100SinusoidalPositionalEmbedding
[ -0.00016463026986457407, 0.032454170286655426, 0.014987517148256302, 0.006817630026489496, -0.0010071498109027743, -0.0005951339844614267, 0.043722979724407196, -0.010029240511357784, 0.005690748803317547, -0.010479993186891079, 0.025692885741591454, 0.023551812395453453, -0.0009508057846687...
[ "False", "ModelSinusoidalPositionalEmbedding", "Module", "None", "__init__", "arange", "bsz", "cat", "class", "contiguous", "cos", "create_position_ids_from_input_ids", "create_position_ids_from_inputs_embeds", "cumsum", "def", "detach", "device", "dim", "dtype", "else", "emb...
m2m_100/modeling_m2m_100.py:eager_attention_forward
[ 0.00002095530362566933, 0.025862814858555794, 0.0269921962171793, -0.01146321278065443, 0.00008955634984886274, 0.03704368323087692, 0.060083046555519104, -0.023942869156599045, 0.021458230912685394, 0.014569009654223919, 0.023152301087975502, 0.026879258453845978, 0.003063444746658206, -0...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "shape", "size", "softmax", "tor...
m2m_100/modeling_m2m_100.py:M2M100Attention
[ -0.0001932329178089276, 0.04201360046863556, 0.04201360046863556, -0.006776387337595224, -0.0007764610345475376, 0.02010328322649002, 0.029138466343283653, -0.028234947472810745, 0.0020611512009054422, 0.014908052049577236, 0.008188134990632534, 0.009261062368750572, -0.0016235094517469406, ...
[ "ALL_ATTENTION_FUNCTIONS", "EncoderDecoderCache", "False", "Instantiating", "Linear", "ModelAttention", "Module", "None", "Please", "True", "ValueError", "__class__", "__init__", "__name__", "_attn_implementation", "a", "and", "attention_interface", "attention_mask", "attn_outp...
m2m_100/modeling_m2m_100.py:M2M100EncoderLayer
[ -0.00013405329082161188, 0.04131663590669632, 0.03793002665042877, 0.008410080336034298, -0.0005714903818443418, 0.03183412924408913, 0.01986810937523842, -0.010216272436082363, 0.002695176750421524, 0.015465517528355122, 0.009200289845466614, 0.01862635277211666, 0.0018061917508020997, 0....
[ "ACT2FN", "False", "GradientCheckpointingLayer", "LayerNorm", "Linear", "ModelAttention", "ModelEncoderLayer", "__init__", "activation_dropout", "activation_fn", "activation_function", "attention_dropout", "attention_mask", "attn_weights", "clamp", "clamp_value", "class", "config",...
m2m_100/modeling_m2m_100.py:M2M100DecoderLayer
[ -0.00016973177844192833, 0.052051082253456116, 0.02217828668653965, -0.007637930102646351, -0.0007779373554512858, 0.03032541275024414, 0.039830390363931656, -0.0341726653277874, 0.005261685233563185, -0.0035502230748534203, -0.0030976049602031708, 0.01900995895266533, -0.0000160228573804488...
[ "ACT2FN", "False", "GradientCheckpointingLayer", "LayerNorm", "Linear", "ModelAttention", "ModelDecoderLayer", "None", "True", "__init__", "activation_dropout", "activation_fn", "activation_function", "attention_dropout", "attention_mask", "cache_position", "class", "config", "cr...
m2m_100/modeling_m2m_100.py:M2M100PreTrainedModel
[ -0.00023800256894901395, 0.04887933284044266, 0.011310450732707977, 0.011253613978624344, -0.0012077742721885443, 0.016027875244617462, 0.030919020995497704, -0.013356562703847885, -0.0036659501492977142, 0.005712061654776335, -0.0064793536439538, 0.0033675588201731443, -0.003552277106791734...
[ "False", "ModelConfig", "ModelDecoderLayer", "ModelEncoderLayer", "ModelPreTrainedModel", "ModelSinusoidalPositionalEmbedding", "PreTrainedModel", "True", "_can_compile_fullgraph", "_init_weights", "_no_split_modules", "_supports_flash_attn", "_supports_flex_attn", "_supports_sdpa", "bas...
m2m_100/modeling_m2m_100.py:M2M100Encoder
[ -0.00013810688687954098, 0.04931735619902611, 0.020267406478524208, 0.00866994634270668, -0.0007353972177952528, 0.04143558442592621, 0.016664311289787292, -0.02105558291077614, 0.002772693755105138, -0.00827585719525814, 0.01801547221839428, 0.02274453267455101, -0.000008081891792244278, ...
[ "BaseModelOutput", "False", "LayerNorm", "ModelEncoder", "ModelEncoderLayer", "ModelPreTrainedModel", "ModelScaledWordEmbedding", "ModelSinusoidalPositionalEmbedding", "ModuleList", "None", "_", "__init__", "all_attentions", "and", "attention_mask", "attentions", "class", "config",...
m2m_100/modeling_m2m_100.py:M2M100Decoder
[ -0.00018160064064431936, 0.07438362389802933, 0.005556093994528055, -0.021770816296339035, -0.0011126361787319183, 0.023811830207705498, 0.049891453236341476, -0.03628469258546829, 0.006349821574985981, -0.012869727797806263, 0, 0.020410140976309776, -0.002409530570730567, 0.01972980238497...
[ "BaseModelOutputWithPastAndCrossAttentions", "DynamicCache", "EncoderDecoderCache", "False", "LayerNorm", "ModelDecoder", "ModelDecoderLayer", "ModelPreTrainedModel", "ModelScaledWordEmbedding", "ModelSinusoidalPositionalEmbedding", "ModuleList", "None", "Setting", "True", "ValueError", ...
m2m_100/modeling_m2m_100.py:M2M100Model
[ -0.000052910811064066365, 0.056867316365242004, 0.009459228254854679, -0.018694570288062096, -0.00009051684173755348, 0.03738914057612419, 0.02529923804104328, -0.040971335023641586, 0.00912339799106121, -0.005988979246467352, 0.02910531871020794, 0.0039739953354001045, -0.001749117742292583...
[ "BaseModelOutput", "ModelDecoder", "ModelEncoder", "ModelModel", "ModelPreTrainedModel", "ModelScaledWordEmbedding", "None", "Seq2SeqModelOutput", "__init__", "_tied_weights_keys", "and", "attention_mask", "attentions", "auto_docstring", "cache_position", "class", "config", "cross_...
m2m_100/modeling_m2m_100.py:M2M100ForConditionalGeneration
[ -0.00024515384575352073, 0.03341147303581238, 0.0030758872162550688, -0.005164104048162699, -0.0009100675815716386, 0.04266735166311264, 0.013883821666240692, -0.02189805917441845, 0, 0.003950681071728468, 0.032959967851638794, -0.0006631499854847789, 0.00034568458795547485, 0.009086566045...
[ "CrossEntropyLoss", "GenerationMixin", "Linear", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "Seq2SeqLMOutput", "__init__", "_tied_weights_keys", "attention_mask", "auto_docstring", "base_model_prefix", "cache_position", "class", "config", "cross_at...
xcodec/modeling_xcodec.py:XcodecOutput
[ -0.000046441651647910476, 0.03374229371547699, -0.02185322903096676, 0.014493334107100964, -0.0000995178270386532, 0.019701875746250153, 0.07065500319004059, -0.042574167251586914, 0.011945676989853382, -0.015852084383368492, 0.02264583483338356, 0.028307292610406876, -0.00653898436576128, ...
[ "ModelOutput", "None", "audio_codes", "audio_values", "class" ]
xcodec/modeling_xcodec.py:XcodecEncoderOutput
[ -0.00004352082032710314, 0.022056439891457558, 0.00046304386341944337, 0.011876544915139675, -0.00027924019377678633, 0.027146387845277786, 0.060626935213804245, -0.055650096386671066, 0.010745445266366005, -0.020925341174006462, 0.017192712053656578, 0.036647625267505646, -0.008313581347465...
[ "ModelEncoderOutput", "ModelOutput", "None", "audio_codes", "class" ]
xcodec/modeling_xcodec.py:XcodecDecoderOutput
[ -0.0001151128817582503, 0.04715023562312126, -0.0068288506008684635, 0.008500644005835056, -0.0008854836924001575, 0.015187816694378853, 0.07480566203594208, -0.06075126677751541, 0.007650579325854778, -0.021874990314245224, 0.010654140263795853, 0.025275247171521187, -0.006913856603205204, ...
[ "ModelDecoderOutput", "ModelOutput", "None", "audio_values", "class" ]
xcodec/modeling_xcodec.py:ResidualUnit
[ -0.00011804773384938017, 0.012213058769702911, 0.00999766681343317, -0.0036355152260512114, -0.00009142042836174369, 0.060440439730882645, 0.017950356006622314, -0.047034479677677155, 0.010111276991665363, 0.012667498551309109, 0.036127932369709015, -0.03340129554271698, 0.004175161942839622...
[ "Conv1d", "ELU", "ModelUnit", "Module", "__init__", "activation", "class", "config", "conv1", "conv2", "def", "dilation", "forward", "groups", "hidden_state", "in_channels", "kernel_size", "nn", "out_channels", "output_tensor", "padding", "return", "self", "stride", "...
xcodec/modeling_xcodec.py:SemanticEncoderBlock
[ -0.0001445545203750953, -0.019645139575004578, 0.00810933019965887, -0.023071616888046265, -0.0005710796103812754, 0.05984914302825928, 0.011307376436889172, -0.05162559822201729, 0.012792183086276054, -0.0015062225284054875, 0.01895984262228012, -0.02718338929116726, 0.004768515005707741, ...
[ "Conv1d", "ModelEncoderBlock", "Module", "ModuleList", "ResidualUnit", "__init__", "block_dilations", "class", "config", "conv", "def", "dilation", "else", "for", "forward", "hidden_state", "if", "in", "in_channels", "kernel", "kernel_size", "nn", "out_channels", "paddi...
xcodec/modeling_xcodec.py:SemanticEncoder
[ -0.00003513231058605015, 0.0070989360101521015, 0.019684700295329094, -0.00639187078922987, 0.00009987298108171672, 0.04276331141591072, -0.0017464514821767807, -0.04570470377802849, 0.01176556758582592, 0.005967631470412016, 0.034844182431697845, -0.02138165570795536, 0.003139370121061802, ...
[ "Conv1d", "ModelEncoder", "ModelEncoderBlock", "Model_hidden_size", "Module", "ModuleList", "Number", "ValueError", "__init__", "block", "channel_ratios", "class", "config", "conv", "conv_blocks", "def", "enumerate", "for", "forward", "hidden_state", "i", "if", "in", "i...
xcodec/modeling_xcodec.py:SemanticDecoderBlock
[ -0.00021795977954752743, 0.020413793623447418, 0.013382376171648502, -0.03243524953722954, -0.0007867816020734608, 0.03969348594546318, 0.017691954970359802, -0.07348965853452682, 0.0069463602267205715, -0.009413027204573154, 0.010830651968717575, -0.022114943712949753, 0.001183716463856399,...
[ "Conv1d", "ConvTranspose1d", "ModelDecoderBlock", "Module", "ModuleList", "ResidualUnit", "__init__", "block_dilations", "class", "config", "conv", "def", "dilation", "else", "for", "forward", "hidden_state", "if", "in", "in_channels", "kernel_size", "nn", "out_channels",...
xcodec/modeling_xcodec.py:SemanticDecoder
[ -0.00015567494847346097, 0.03940247371792793, 0.008274518884718418, -0.011595584452152252, -0.0004749405197799206, 0.03107166290283203, 0.02589305303990841, -0.06574583798646927, 0.00917514692991972, -0.013734576292335987, 0.013565707951784134, -0.02454211190342903, -0.0006121455226093531, ...
[ "Conv1d", "ModelDecoder", "ModelDecoderBlock", "Model_hidden_size", "Module", "ModuleList", "__init__", "block", "channel_ratios", "class", "config", "conv1", "conv2", "conv_blocks", "def", "else", "enumerate", "for", "forward", "hidden_state", "i", "if", "in", "in_chan...
xcodec/modeling_xcodec.py:XcodecEuclideanCodebook
[ -0.00022551401343662292, 0.006446313578635454, 0.018369141966104507, 0.00550503795966506, -0.0009626685059629381, 0.006931213196367025, 0.045181240886449814, -0.036966472864151, 0.007244972046464682, 0.00036010934854857624, 0.0353691540658474, 0.002681210171431303, -0.0016115783946588635, ...
[ "F", "ModelEuclideanCodebook", "Module", "Tensor", "True", "__init__", "class", "clone", "cluster_size", "codebook_dim", "codebook_size", "config", "decode", "def", "dim", "dist", "embed", "embed_avg", "embed_ind", "embedding", "encode", "hidden_states", "indices", "ini...
xcodec/modeling_xcodec.py:XcodecVectorQuantization
[ -0.0001221633137902245, 0.018449297174811363, 0.017099348828196526, 0.006159140262752771, -0.000379673030693084, 0.009505887515842915, 0.0431983545422554, -0.04139842465519905, 0.002235852414742112, 0.017774323001503944, 0.0431983545422554, 0.009224648587405682, -0.003346747485920787, 0.00...
[ "ModelEuclideanCodebook", "ModelVectorQuantization", "Module", "__init__", "class", "codebook", "config", "decode", "def", "embed_in", "embed_ind", "encode", "hidden_states", "nn", "permute", "quantize", "return", "self", "super" ]
xcodec/modeling_xcodec.py:XcodecResidualVectorQuantization
[ -0.00011209374497411773, 0.01124114915728569, -0.011975495144724846, -0.003036240115761757, -0.00028773670783266425, 0.043382927775382996, 0.02632349543273449, -0.043608881533145905, 0.00610072398558259, -0.0036717320326715708, 0.029373856261372566, 0.012709842063486576, 0.000476619083201512...
[ "ModelResidualVectorQuantization", "ModelVectorQuantization", "Module", "ModuleList", "None", "_", "__init__", "all_indices", "and", "append", "bandwidth", "bw_per_q", "class", "codebook_size", "codes", "config", "decode", "def", "device", "embeddings", "encode", "enumerate...
xcodec/modeling_xcodec.py:XcodecPreTrainedModel
[ -0.00004884477675659582, 0.04956846684217453, -0.010597783140838146, -0.012111752294003963, -0.00031190566369332373, 0.01626114919781685, 0.020746983587741852, -0.043288301676511765, 0.007625918369740248, -0.016934024170041084, 0.028597192838788033, 0.023887068033218384, -0.00080955290468409...
[ "AttributeError", "Conv1d", "ConvTranspose1d", "Embedding", "GroupNorm", "LayerNorm", "Linear", "Model", "ModelConfig", "ModelEuclideanCodebook", "ModelModel", "ModelPreTrainedModel", "None", "PreTrainedAudioTokenizerBase", "Snake1d", "Tensor", "True", "ValueError", "__class__", ...
xcodec/modeling_xcodec.py:XcodecModel
[ -0.00010499545896891505, 0.046529751271009445, -0.01750512421131134, -0.008074944838881493, -0.0003988062671851367, 0.0313962884247303, 0.031848032027482986, -0.032977394759655, 0.006126793567091227, -0.022700194269418716, 0.0478849858045578, 0.024394238367676735, -0.003811599686741829, -0...
[ "Audio", "AutoModel", "ConvTranspose1d", "F", "False", "Identity", "Linear", "ModelDecoderOutput", "ModelEncoderOutput", "ModelModel", "ModelOutput", "ModelPreTrainedModel", "ModelResidualVectorQuantization", "None", "Select", "SemanticDecoder", "SemanticEncoder", "Tanh", "This",...
qwen3_vl/modeling_qwen3_vl.py:BaseModelOutputWithDeepstackFeatures
[ -0.000323236221447587, 0.014142042025923729, 0.015544557943940163, 0.02173900604248047, -0.001380602247081697, 0.05095810070633888, 0.0631132423877716, -0.030621610581874847, 0.00876572821289301, -0.011629199609160423, 0.031322870403528214, 0.01601206324994564, -0.0004565483541227877, 0.01...
[ "ModelModelOutputWithDeepstackFeatures", "ModelModelOutputWithPooling", "None", "class", "deepstack_features", "r" ]
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionMLP
[ -0.00020412541925907135, 0.02817285805940628, 0.032262466847896576, 0.033852867782115936, -0.00048280099872499704, 0.04589449241757393, 0.025673652067780495, -0.030444862321019173, 0.0004544009279925376, 0.005850411951541901, 0.04589449241757393, -0.022152045741677284, 0.0011999025009572506,...
[ "ACT2FN", "Linear", "ModelVisionMLP", "Module", "__init__", "act_fn", "class", "config", "def", "forward", "hidden_act", "hidden_size", "hidden_state", "intermediate_size", "linear_fc1", "linear_fc2", "nn", "return", "self", "super" ]
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionPatchEmbed
[ -0.00006803657743148506, 0.010169493034482002, 0.01101226918399334, 0.00854012742638588, -0.00019313610391691327, 0.008989606983959675, 0.027081191539764404, -0.03191310539841652, 0.013877706602215767, -0.0005021538236178458, 0.008483941666781902, -0.00022386229829862714, -0.0015661581419408...
[ "Conv3d", "ModelVisionPatchEmbed", "Module", "__init__", "class", "config", "def", "dtype", "embed_dim", "forward", "hidden_size", "hidden_states", "in_channels", "kernel_size", "nn", "patch_size", "proj", "return", "self", "stride", "super", "target_dtype", "temporal_pat...
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionRotaryEmbedding
[ -0.00022179921506904066, 0.03434310480952263, 0.01224904041737318, 0.008013390935957432, -0.0010374479461461306, 0.04464603587985039, 0.0668545737862587, -0.016713643446564674, 0.005323181394487619, 0.0345720574259758, 0.00887196883559227, -0.00875749159604311, -0.0022466115187853575, 0.00...
[ "False", "ModelVisionRotaryEmbedding", "Module", "Tensor", "__init__", "arange", "class", "def", "device", "dim", "dtype", "float", "forward", "freqs", "inv_freq", "nn", "outer", "persistent", "register_buffer", "return", "self", "seq", "seqlen", "super", "theta", "...
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionPatchMerger
[ -0.00025101262144744396, 0.03235587850213051, 0.011313244700431824, 0.031224554404616356, -0.0006858654669485986, 0.021042635664343834, 0.00933342706412077, -0.04683683067560196, 0.00574147142469883, 0.0191193837672472, 0.011200112290680408, 0.007212193217128515, -0.000643440755084157, -0....
[ "False", "GELU", "LayerNorm", "Linear", "ModelVisionPatchMerger", "Module", "__init__", "act_fn", "class", "config", "def", "else", "eps", "forward", "hidden_size", "if", "linear_fc1", "linear_fc2", "nn", "norm", "out_hidden_size", "return", "self", "spatial_merge_size"...
qwen3_vl/modeling_qwen3_vl.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
qwen3_vl/modeling_qwen3_vl.py:apply_rotary_pos_emb_vision
[ -0.00006463831959990785, 0.03513491153717041, 0.018587501719594002, -0.0022809358779340982, -0.0004639792023226619, 0.01700076460838318, 0.060749396681785583, -0.009180412627756596, 0.016887426376342773, 0.04692211002111435, 0.009010405279695988, 0.01983422413468361, 0.0013458938337862492, ...
[ "Model_rotary_pos_emb_vision", "cos", "def", "dtype", "float", "k", "k_embed", "orig_q_dtype", "q", "q_embed", "return", "rotate_half", "sin", "unsqueeze" ]
qwen3_vl/modeling_qwen3_vl.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
qwen3_vl/modeling_qwen3_vl.py:eager_attention_forward
[ 0, 0.021594731137156487, 0.01775064319372177, -0.018768195062875748, -0.00008832923776935786, 0.039797618985176086, 0.05359111353754997, -0.035049039870500565, 0.02069023996591568, 0.011645326390862465, 0.029395969584584236, 0.022499222308397293, 0.002741739386692643, -0.014584923163056374...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "causal_mask", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "key_states", "kwargs", "matmul", "module", "nn", "not", "num_key_value_groups", ...
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionAttention
[ -0.00014312515850178897, 0.03664004057645798, 0.021825702860951424, 0.021825702860951424, -0.0004081717343069613, 0.0252182986587286, 0.03958028927445412, -0.029628673568367958, 0.0055129691027104855, 0.03189040347933769, 0.028837068006396294, 0.03189040347933769, 0.0010036430321633816, -0...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "ModelVisionAttention", "Module", "None", "Tensor", "_", "__init__", "_attn_implementation", "apply_rotary_pos_emb_vision", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_outputs", "cat", "class", "...
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionBlock
[ -0.00016528498963452876, 0.013067236170172691, 0.01606534607708454, 0.01855434477329254, -0.00040481562609784305, 0.043896861374378204, 0.026587016880512238, -0.01866747997701168, 0.005232551135122776, 0.021043341606855392, 0.021043341606855392, 0.019120024517178535, 0.0016475465381518006, ...
[ "GradientCheckpointingLayer", "LayerNorm", "ModelVisionAttention", "ModelVisionBlock", "ModelVisionMLP", "None", "Tensor", "__init__", "attn", "attn_implementation", "class", "config", "cu_seqlens", "def", "eps", "forward", "hidden_size", "hidden_states", "kwargs", "mlp", "nn...
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextRotaryEmbedding
[ -0.00028448071680031717, 0.04476313292980194, 0.002913662465289235, -0.02087399922311306, -0.0013263686560094357, 0.05728752911090851, 0.032470665872097015, -0.017163066193461418, -0.007769766263663769, 0.026440398767590523, 0.02052609995007515, 0.0017250040546059608, -0.0010509478161111474,...
[ "False", "ModelTextRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "apply_interleaved_mrope", "arange", "attention_factor", "attention_scaling", "base", "cat", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cpu", ...
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextRMSNorm
[ -0.0000997777096927166, 0.04746946692466736, 0.03865370899438858, 0.048825737088918686, -0.0005686446093022823, 0.03435885161161423, 0.027125408872961998, -0.03481094166636467, 0.008702735416591167, 0.035715121775865555, 0.021474281325936317, 0.011415275745093822, 0.002090916968882084, 0.0...
[ "ModelTextRMSNorm", "Module", "Parameter", "True", "__init__", "class", "def", "dtype", "eps", "extra_repr", "f", "float32", "forward", "hidden_size", "hidden_states", "input_dtype", "keepdim", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "shape", "super"...
qwen3_vl/modeling_qwen3_vl.py:apply_rotary_pos_emb
[ -0.0001444592053303495, 0.027112245559692383, 0.028019767254590988, 0.0028360087890177965, -0.0005707467789761722, 0.021667107939720154, 0.046510547399520874, -0.002183726755902171, 0.01293220091611147, 0.03652779385447502, 0.007884104736149311, 0.0017654155381023884, -0.0007834474672563374,...
[ "Model_rotary_pos_emb", "cos", "def", "k", "k_embed", "q", "q_embed", "return", "rotate_half", "sin", "unsqueeze", "unsqueeze_dim" ]
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextAttention
[ -0.00005613891698885709, 0.046707578003406525, 0.0357043519616127, 0.0008385750697925687, -0.00038420071359723806, 0.025936178863048553, 0.033234238624572754, -0.027171235531568527, 0.0009052400127984583, 0.01998545415699482, 0.022455567494034767, 0.037276241928339005, -0.0011087435996159911...
[ "ALL_ATTENTION_FUNCTIONS", "Linear", "ModelTextAttention", "ModelTextRMSNorm", "Module", "None", "Tensor", "True", "__init__", "_attn_implementation", "apply_rotary_pos_emb", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "cache_kwargs", ...
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextMLP
[ -0.00019116139446850866, 0.03247242420911789, 0.03338714316487312, 0.01726526953279972, -0.0005788438720628619, 0.057398442178964615, 0.02641243115067482, -0.016922250390052795, -0.002029526513069868, -0.008232446387410164, 0.035216573625802994, -0.026069412007927895, -0.0016507769469171762,...
[ "ACT2FN", "Linear", "ModelTextMLP", "Module", "__init__", "act_fn", "class", "config", "def", "down_proj", "forward", "gate_proj", "hidden_act", "hidden_size", "intermediate_size", "nn", "return", "self", "super", "up_proj", "x" ]
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextDecoderLayer
[ -0.00016621610848233104, 0.0436830073595047, 0.024897051975131035, 0.0035789511166512966, -0.0007568137953057885, 0.039382610470056534, 0.036666568368673325, -0.03825092315673828, 0.005573544651269913, -0.0032252997625619173, 0.01007198914885521, 0.028518440201878548, -0.002178492024540901, ...
[ "False", "GradientCheckpointingLayer", "ModelTextAttention", "ModelTextDecoderLayer", "ModelTextMLP", "ModelTextRMSNorm", "None", "Tensor", "_", "__init__", "attention_mask", "cache_position", "class", "config", "def", "eps", "forward", "hidden_size", "hidden_states", "input_la...
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLModelOutputWithPast
[ -0.0002648343506734818, 0.008071142248809338, 0.014355102553963661, 0.0052174171432852745, -0.0014196563279256225, 0.042200542986392975, 0.04911866411566734, -0.03851087763905525, 0.016949398443102837, 0.0025078190956264734, 0.008071142248809338, 0.03182335942983627, -0.001830419758334756, ...
[ "ModelModelOutputWithPast", "ModelOutput", "None", "attentions", "class", "hidden_states", "last_hidden_state", "past_key_values", "r", "rope_deltas" ]
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLPreTrainedModel
[ -0.0003329527680762112, 0.04890584945678711, 0.004803253337740898, 0.010014055296778679, -0.0017757480964064598, 0.032836783677339554, 0.02724754437804222, -0.034234095364809036, -0.0043083722703158855, 0.015719737857580185, 0.009431842714548111, -0.0039881556294858456, -0.005560129415243864...
[ "ModelConfig", "ModelPreTrainedModel", "ModelTextAttention", "ModelTextDecoderLayer", "ModelVisionBlock", "ModelVisionRotaryEmbedding", "PreTrainedModel", "True", "_can_compile_fullgraph", "_can_record_outputs", "_init_weights", "_no_split_modules", "_skip_keys_device_placement", "_support...
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionModel
[ -0.0002279315231135115, 0.01696234568953514, -0.023521119728684425, 0.01933707483112812, -0.000816312909591943, 0.03460318595170975, 0.047494567930698395, -0.05156553164124489, 0.0019223992712795734, 0.02159872092306614, 0.01809316873550415, 0.00689802085980773, -0.0006749600288458169, 0.0...
[ "BaseModelOutputWithDeepstackFeatures", "Embedding", "F", "False", "ModelPreTrainedModel", "ModelVisionAttention", "ModelVisionBlock", "ModelVisionConfig", "ModelVisionModel", "ModelVisionPatchEmbed", "ModelVisionPatchMerger", "ModelVisionRotaryEmbedding", "ModuleList", "None", "T", "T...
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextModel
[ -0.00009935703565133736, 0.04078628122806549, 0.0032774689607322216, 0.002044916618615389, -0.0006582950591109693, 0.03159816190600395, 0.03854527696967125, -0.04190678521990776, 0.004874184727668762, 0.008291716687381268, 0.03159816190600395, 0.020505191758275032, -0.002212991937994957, 0...
[ "BaseModelOutputWithPast", "DynamicCache", "Embedding", "False", "ModelPreTrainedModel", "ModelTextConfig", "ModelTextDecoderLayer", "ModelTextModel", "ModelTextRMSNorm", "ModelTextRotaryEmbedding", "ModuleList", "None", "ValueError", "You", "__init__", "_deepstack_process", "_no_spl...
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLModel
[ -0.00008527900354238227, 0.03896069899201393, -0.018472744151949883, 0.0014134448720142245, -0.0003113777202088386, 0.05396280437707901, 0.03269116207957268, -0.0470215305685997, 0.005877691321074963, 0.022055337205529213, 0.033586807548999786, 0.020264040678739548, 0.0004110885492991656, ...
[ "False", "Image", "ModelConfig", "ModelModel", "ModelModelOutputWithPast", "ModelPreTrainedModel", "ModelTextDecoderLayer", "ModelTextModel", "ModelVisionBlock", "ModelVisionModel", "None", "True", "ValueError", "Video", "You", "_", "__init__", "_checkpoint_conversion_mapping", "...
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLCausalLMOutputWithPast
[ -0.0002988479973282665, 0.017503436654806137, 0.020285440608859062, -0.0002644352207425982, -0.0015503872418776155, 0.049844224005937576, 0.049612391740083694, -0.02874736674129963, 0.016923854127526283, 0.0007607040461152792, 0.017155686393380165, 0.02156052552163601, -0.0010577408829703927...
[ "ModelCausalLMOutputWithPast", "ModelOutput", "None", "attentions", "class", "hidden_states", "logits", "loss", "past_key_values", "r", "rope_deltas" ]
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLForConditionalGeneration
[ -0.00028949574334546924, 0.03455477952957153, -0.010002699680626392, 0.019778065383434296, -0.000994586618617177, 0.03910146281123161, 0.031826771795749664, -0.050922833383083344, -0.0006997627206146717, 0.01182137243449688, 0.03523678332567215, 0.00115087884478271, 0.0005896477960050106, ...
[ "False", "GenerationMixin", "Linear", "ModelCausalLMOutputWithPast", "ModelConfig", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "Tensor", "True", "__init__", "_checkpoint_conversion_mapping", "_expand_dict_for_generation", "_expand_dict_for_generation_v...
falcon_mamba/modeling_falcon_mamba.py:FalconMambaCache
[ -0.00018014491070061922, 0.024301903322339058, 0.0095512131229043, -0.014581141993403435, -0.0004415316507220268, 0.006697152275592089, -0.013733400963246822, -0.05222083255648613, 0.006188507657498121, 0.031648989766836166, 0.012998691760003567, -0.0025573514867573977, 0.0011656435672193766...
[ "ModelCache", "None", "True", "_", "__init__", "_dtype", "_dynamo", "append", "cache_position", "clamp", "class", "config", "conv_kernel", "conv_kernel_size", "conv_state", "conv_states", "def", "device", "dims", "dtype", "else", "float16", "for", "if", "in", "inter...
falcon_mamba/modeling_falcon_mamba.py:rms_forward
[ -0.00017206810298375785, 0.03139933943748474, 0.039757438004016876, 0.03907975181937218, -0.00045884825522080064, 0.007510993164032698, 0.029592182487249374, -0.032980602234601974, 0.01248067244887352, 0.032980602234601974, 0.03456186130642891, -0.0034166546538472176, 0.0024142477195709944, ...
[ "Model_forward", "True", "def", "float32", "hidden_states", "keepdim", "mean", "pow", "return", "rsqrt", "to", "torch", "variance", "variance_epsilon" ]
falcon_mamba/modeling_falcon_mamba.py:FalconMambaMixer
[ -0.0001887798571260646, 0.022454118356108665, 0.00946036446839571, 0.0008370428113266826, -0.00034550277632661164, 0.05015132948756218, 0.013905595988035202, -0.035333890467882156, 0.008662502281367779, 0.010144246742129326, 0.020174512639641762, -0.0016028479440137744, 0.001937665045261383,...
[ "A", "ACT2FN", "A_log", "B", "C", "Conv1d", "D", "Falling", "False", "ImportError", "Linear", "Mamba", "Model", "ModelMixer", "Model_inner_fn", "Model_ssm", "Module", "None", "Parameter", "The", "To", "True", "_", "__init__", "_is_quantized", "act", "activation", ...
falcon_mamba/modeling_falcon_mamba.py:FalconMambaRMSNorm
[ -0.00008794592577032745, 0.03575112298130989, 0.02059083804488182, 0.05475805327296257, -0.00028284118161536753, 0.044802043586969376, 0.02138279192149639, -0.028962936252355576, 0.008654939942061901, 0.0366562157869339, 0.014481468126177788, 0.0026304228231310844, 0.0032385315280407667, 0...
[ "ModelRMSNorm", "Module", "Parameter", "__init__", "class", "def", "device", "eps", "extra_repr", "f", "forward", "hidden_size", "hidden_states", "nn", "ones", "return", "rms_forward", "self", "shape", "super", "to", "torch", "variance_epsilon", "weight" ]
falcon_mamba/modeling_falcon_mamba.py:FalconMambaBlock
[ -0.00018491271475795656, -0.0019252155907452106, 0.02095087617635727, 0.004699791315943003, -0.0003945984353777021, 0.05096159130334854, 0.02672652341425419, -0.03623935207724571, 0.004190175328403711, 0.015401724725961685, -0.005350966937839985, -0.007587614469230175, 0.000422910408815369, ...
[ "GradientCheckpointingLayer", "ModelBlock", "ModelMixer", "ModelRMSNorm", "None", "__init__", "attention_mask", "cache_params", "cache_position", "class", "config", "def", "dtype", "eps", "float32", "forward", "hidden_size", "hidden_states", "if", "layer_idx", "layer_norm_eps...
falcon_mamba/modeling_falcon_mamba.py:FalconMambaPreTrainedModel
[ -0.00028834445402026176, 0.028934119269251823, -0.004471118561923504, 0.0049837310798466206, -0.0011818562634289265, 0.0323515348136425, 0.011505299247801304, -0.003146870294585824, -0.009170065633952618, 0.020846234634518623, -0.00032750231912359595, -0.03349067270755768, -0.002748171798884...
[ "A", "A_log", "D", "Embedding", "Linear", "ModelBlock", "ModelConfig", "ModelMixer", "ModelPreTrainedModel", "ModelRMSNorm", "None", "PreTrainedModel", "True", "_init_weights", "_is_stateful", "_no_split_modules", "a", "arange", "b_c_rms", "backbone", "base_model_prefix", "...
falcon_mamba/modeling_falcon_mamba.py:FalconMambaOutput
[ -0.00017384765669703484, 0.02076900005340576, 0.031952306628227234, 0.02179603837430477, -0.001112625002861023, 0.03309345990419388, 0.05135192349553108, -0.01997019164264202, 0.017916115000844002, -0.016204385086894035, -0.003166701877489686, 0.007988076657056808, -0.0008736958843655884, ...
[ "ModelOutput", "None", "cache_params", "class", "hidden_states", "last_hidden_state", "r" ]
falcon_mamba/modeling_falcon_mamba.py:FalconMambaCausalLMOutput
[ -0.0001988249714486301, 0.028416814282536507, 0.035378362983465195, 0.004079923965036869, -0.0011483702110126615, 0.03857382759451866, 0.050899192690849304, -0.007817476987838745, 0.01791742816567421, -0.015977324917912483, 0.0023538023233413696, -0.0003940835886169225, -0.002425129758194089...
[ "ModelCausalLMOutput", "ModelOutput", "None", "cache_params", "class", "hidden_states", "logits", "loss", "r" ]
falcon_mamba/modeling_falcon_mamba.py:FalconMambaModel
[ 0.000011523719877004623, 0.0359627865254879, -0.0007340061129070818, 0.00797924306243658, 0.00009877474076347426, 0.040682900696992874, 0.027758775278925896, 0.0007550780428573489, 0.005310130305588245, 0.0076420921832323074, 0.012980318628251553, -0.013879387639462948, 0.0006497183348983526...
[ "Embedding", "False", "ModelBlock", "ModelCache", "ModelModel", "ModelOutput", "ModelPreTrainedModel", "ModelRMSNorm", "ModuleList", "None", "True", "ValueError", "You", "__init__", "a", "all_hidden_states", "and", "arange", "are", "attention_mask", "auto_docstring", "autom...
falcon_mamba/modeling_falcon_mamba.py:FalconMambaForCausalLM
[ -0.00022307969629764557, 0.030140545219182968, 0.014617031440138817, -0.00611875718459487, -0.0008994006784632802, 0.019602686166763306, 0.012747411616146564, -0.007648446597158909, -0.0009418920380994678, 0.007875067181885242, 0.012917377054691315, -0.013200652785599232, 0.00167840917129069...
[ "Any", "CrossEntropyLoss", "False", "GenerationMixin", "Linear", "ModelCache", "ModelCausalLMOutput", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "Model_outputs", "None", "True", "__init__", "_tied_weights_keys", "_update_model_kwargs_for_generation", "and", "arange",...