identifier stringlengths 24 102 | embedding listlengths 2.56k 2.56k | tokens listlengths 4 448 |
|---|---|---|
m2m_100/modeling_m2m_100.py:M2M100SinusoidalPositionalEmbedding | [
-0.00016463026986457407,
0.032454170286655426,
0.014987517148256302,
0.006817630026489496,
-0.0010071498109027743,
-0.0005951339844614267,
0.043722979724407196,
-0.010029240511357784,
0.005690748803317547,
-0.010479993186891079,
0.025692885741591454,
0.023551812395453453,
-0.0009508057846687... | [
"False",
"ModelSinusoidalPositionalEmbedding",
"Module",
"None",
"__init__",
"arange",
"bsz",
"cat",
"class",
"contiguous",
"cos",
"create_position_ids_from_input_ids",
"create_position_ids_from_inputs_embeds",
"cumsum",
"def",
"detach",
"device",
"dim",
"dtype",
"else",
"emb... |
m2m_100/modeling_m2m_100.py:eager_attention_forward | [
0.00002095530362566933,
0.025862814858555794,
0.0269921962171793,
-0.01146321278065443,
0.00008955634984886274,
0.03704368323087692,
0.060083046555519104,
-0.023942869156599045,
0.021458230912685394,
0.014569009654223919,
0.023152301087975502,
0.026879258453845978,
0.003063444746658206,
-0... | [
"Model_attention_forward",
"None",
"attention_mask",
"attn_output",
"attn_weights",
"contiguous",
"def",
"dim",
"dropout",
"functional",
"if",
"is",
"key",
"kwargs",
"matmul",
"module",
"nn",
"not",
"p",
"query",
"return",
"scaling",
"shape",
"size",
"softmax",
"tor... |
m2m_100/modeling_m2m_100.py:M2M100Attention | [
-0.0001932329178089276,
0.04201360046863556,
0.04201360046863556,
-0.006776387337595224,
-0.0007764610345475376,
0.02010328322649002,
0.029138466343283653,
-0.028234947472810745,
0.0020611512009054422,
0.014908052049577236,
0.008188134990632534,
0.009261062368750572,
-0.0016235094517469406,
... | [
"ALL_ATTENTION_FUNCTIONS",
"EncoderDecoderCache",
"False",
"Instantiating",
"Linear",
"ModelAttention",
"Module",
"None",
"Please",
"True",
"ValueError",
"__class__",
"__init__",
"__name__",
"_attn_implementation",
"a",
"and",
"attention_interface",
"attention_mask",
"attn_outp... |
m2m_100/modeling_m2m_100.py:M2M100EncoderLayer | [
-0.00013405329082161188,
0.04131663590669632,
0.03793002665042877,
0.008410080336034298,
-0.0005714903818443418,
0.03183412924408913,
0.01986810937523842,
-0.010216272436082363,
0.002695176750421524,
0.015465517528355122,
0.009200289845466614,
0.01862635277211666,
0.0018061917508020997,
0.... | [
"ACT2FN",
"False",
"GradientCheckpointingLayer",
"LayerNorm",
"Linear",
"ModelAttention",
"ModelEncoderLayer",
"__init__",
"activation_dropout",
"activation_fn",
"activation_function",
"attention_dropout",
"attention_mask",
"attn_weights",
"clamp",
"clamp_value",
"class",
"config",... |
m2m_100/modeling_m2m_100.py:M2M100DecoderLayer | [
-0.00016973177844192833,
0.052051082253456116,
0.02217828668653965,
-0.007637930102646351,
-0.0007779373554512858,
0.03032541275024414,
0.039830390363931656,
-0.0341726653277874,
0.005261685233563185,
-0.0035502230748534203,
-0.0030976049602031708,
0.01900995895266533,
-0.0000160228573804488... | [
"ACT2FN",
"False",
"GradientCheckpointingLayer",
"LayerNorm",
"Linear",
"ModelAttention",
"ModelDecoderLayer",
"None",
"True",
"__init__",
"activation_dropout",
"activation_fn",
"activation_function",
"attention_dropout",
"attention_mask",
"cache_position",
"class",
"config",
"cr... |
m2m_100/modeling_m2m_100.py:M2M100PreTrainedModel | [
-0.00023800256894901395,
0.04887933284044266,
0.011310450732707977,
0.011253613978624344,
-0.0012077742721885443,
0.016027875244617462,
0.030919020995497704,
-0.013356562703847885,
-0.0036659501492977142,
0.005712061654776335,
-0.0064793536439538,
0.0033675588201731443,
-0.003552277106791734... | [
"False",
"ModelConfig",
"ModelDecoderLayer",
"ModelEncoderLayer",
"ModelPreTrainedModel",
"ModelSinusoidalPositionalEmbedding",
"PreTrainedModel",
"True",
"_can_compile_fullgraph",
"_init_weights",
"_no_split_modules",
"_supports_flash_attn",
"_supports_flex_attn",
"_supports_sdpa",
"bas... |
m2m_100/modeling_m2m_100.py:M2M100Encoder | [
-0.00013810688687954098,
0.04931735619902611,
0.020267406478524208,
0.00866994634270668,
-0.0007353972177952528,
0.04143558442592621,
0.016664311289787292,
-0.02105558291077614,
0.002772693755105138,
-0.00827585719525814,
0.01801547221839428,
0.02274453267455101,
-0.000008081891792244278,
... | [
"BaseModelOutput",
"False",
"LayerNorm",
"ModelEncoder",
"ModelEncoderLayer",
"ModelPreTrainedModel",
"ModelScaledWordEmbedding",
"ModelSinusoidalPositionalEmbedding",
"ModuleList",
"None",
"_",
"__init__",
"all_attentions",
"and",
"attention_mask",
"attentions",
"class",
"config",... |
m2m_100/modeling_m2m_100.py:M2M100Decoder | [
-0.00018160064064431936,
0.07438362389802933,
0.005556093994528055,
-0.021770816296339035,
-0.0011126361787319183,
0.023811830207705498,
0.049891453236341476,
-0.03628469258546829,
0.006349821574985981,
-0.012869727797806263,
0,
0.020410140976309776,
-0.002409530570730567,
0.01972980238497... | [
"BaseModelOutputWithPastAndCrossAttentions",
"DynamicCache",
"EncoderDecoderCache",
"False",
"LayerNorm",
"ModelDecoder",
"ModelDecoderLayer",
"ModelPreTrainedModel",
"ModelScaledWordEmbedding",
"ModelSinusoidalPositionalEmbedding",
"ModuleList",
"None",
"Setting",
"True",
"ValueError",
... |
m2m_100/modeling_m2m_100.py:M2M100Model | [
-0.000052910811064066365,
0.056867316365242004,
0.009459228254854679,
-0.018694570288062096,
-0.00009051684173755348,
0.03738914057612419,
0.02529923804104328,
-0.040971335023641586,
0.00912339799106121,
-0.005988979246467352,
0.02910531871020794,
0.0039739953354001045,
-0.001749117742292583... | [
"BaseModelOutput",
"ModelDecoder",
"ModelEncoder",
"ModelModel",
"ModelPreTrainedModel",
"ModelScaledWordEmbedding",
"None",
"Seq2SeqModelOutput",
"__init__",
"_tied_weights_keys",
"and",
"attention_mask",
"attentions",
"auto_docstring",
"cache_position",
"class",
"config",
"cross_... |
m2m_100/modeling_m2m_100.py:M2M100ForConditionalGeneration | [
-0.00024515384575352073,
0.03341147303581238,
0.0030758872162550688,
-0.005164104048162699,
-0.0009100675815716386,
0.04266735166311264,
0.013883821666240692,
-0.02189805917441845,
0,
0.003950681071728468,
0.032959967851638794,
-0.0006631499854847789,
0.00034568458795547485,
0.009086566045... | [
"CrossEntropyLoss",
"GenerationMixin",
"Linear",
"ModelForConditionalGeneration",
"ModelModel",
"ModelPreTrainedModel",
"None",
"Seq2SeqLMOutput",
"__init__",
"_tied_weights_keys",
"attention_mask",
"auto_docstring",
"base_model_prefix",
"cache_position",
"class",
"config",
"cross_at... |
xcodec/modeling_xcodec.py:XcodecOutput | [
-0.000046441651647910476,
0.03374229371547699,
-0.02185322903096676,
0.014493334107100964,
-0.0000995178270386532,
0.019701875746250153,
0.07065500319004059,
-0.042574167251586914,
0.011945676989853382,
-0.015852084383368492,
0.02264583483338356,
0.028307292610406876,
-0.00653898436576128,
... | [
"ModelOutput",
"None",
"audio_codes",
"audio_values",
"class"
] |
xcodec/modeling_xcodec.py:XcodecEncoderOutput | [
-0.00004352082032710314,
0.022056439891457558,
0.00046304386341944337,
0.011876544915139675,
-0.00027924019377678633,
0.027146387845277786,
0.060626935213804245,
-0.055650096386671066,
0.010745445266366005,
-0.020925341174006462,
0.017192712053656578,
0.036647625267505646,
-0.008313581347465... | [
"ModelEncoderOutput",
"ModelOutput",
"None",
"audio_codes",
"class"
] |
xcodec/modeling_xcodec.py:XcodecDecoderOutput | [
-0.0001151128817582503,
0.04715023562312126,
-0.0068288506008684635,
0.008500644005835056,
-0.0008854836924001575,
0.015187816694378853,
0.07480566203594208,
-0.06075126677751541,
0.007650579325854778,
-0.021874990314245224,
0.010654140263795853,
0.025275247171521187,
-0.006913856603205204,
... | [
"ModelDecoderOutput",
"ModelOutput",
"None",
"audio_values",
"class"
] |
xcodec/modeling_xcodec.py:ResidualUnit | [
-0.00011804773384938017,
0.012213058769702911,
0.00999766681343317,
-0.0036355152260512114,
-0.00009142042836174369,
0.060440439730882645,
0.017950356006622314,
-0.047034479677677155,
0.010111276991665363,
0.012667498551309109,
0.036127932369709015,
-0.03340129554271698,
0.004175161942839622... | [
"Conv1d",
"ELU",
"ModelUnit",
"Module",
"__init__",
"activation",
"class",
"config",
"conv1",
"conv2",
"def",
"dilation",
"forward",
"groups",
"hidden_state",
"in_channels",
"kernel_size",
"nn",
"out_channels",
"output_tensor",
"padding",
"return",
"self",
"stride",
"... |
xcodec/modeling_xcodec.py:SemanticEncoderBlock | [
-0.0001445545203750953,
-0.019645139575004578,
0.00810933019965887,
-0.023071616888046265,
-0.0005710796103812754,
0.05984914302825928,
0.011307376436889172,
-0.05162559822201729,
0.012792183086276054,
-0.0015062225284054875,
0.01895984262228012,
-0.02718338929116726,
0.004768515005707741,
... | [
"Conv1d",
"ModelEncoderBlock",
"Module",
"ModuleList",
"ResidualUnit",
"__init__",
"block_dilations",
"class",
"config",
"conv",
"def",
"dilation",
"else",
"for",
"forward",
"hidden_state",
"if",
"in",
"in_channels",
"kernel",
"kernel_size",
"nn",
"out_channels",
"paddi... |
xcodec/modeling_xcodec.py:SemanticEncoder | [
-0.00003513231058605015,
0.0070989360101521015,
0.019684700295329094,
-0.00639187078922987,
0.00009987298108171672,
0.04276331141591072,
-0.0017464514821767807,
-0.04570470377802849,
0.01176556758582592,
0.005967631470412016,
0.034844182431697845,
-0.02138165570795536,
0.003139370121061802,
... | [
"Conv1d",
"ModelEncoder",
"ModelEncoderBlock",
"Model_hidden_size",
"Module",
"ModuleList",
"Number",
"ValueError",
"__init__",
"block",
"channel_ratios",
"class",
"config",
"conv",
"conv_blocks",
"def",
"enumerate",
"for",
"forward",
"hidden_state",
"i",
"if",
"in",
"i... |
xcodec/modeling_xcodec.py:SemanticDecoderBlock | [
-0.00021795977954752743,
0.020413793623447418,
0.013382376171648502,
-0.03243524953722954,
-0.0007867816020734608,
0.03969348594546318,
0.017691954970359802,
-0.07348965853452682,
0.0069463602267205715,
-0.009413027204573154,
0.010830651968717575,
-0.022114943712949753,
0.001183716463856399,... | [
"Conv1d",
"ConvTranspose1d",
"ModelDecoderBlock",
"Module",
"ModuleList",
"ResidualUnit",
"__init__",
"block_dilations",
"class",
"config",
"conv",
"def",
"dilation",
"else",
"for",
"forward",
"hidden_state",
"if",
"in",
"in_channels",
"kernel_size",
"nn",
"out_channels",... |
xcodec/modeling_xcodec.py:SemanticDecoder | [
-0.00015567494847346097,
0.03940247371792793,
0.008274518884718418,
-0.011595584452152252,
-0.0004749405197799206,
0.03107166290283203,
0.02589305303990841,
-0.06574583798646927,
0.00917514692991972,
-0.013734576292335987,
0.013565707951784134,
-0.02454211190342903,
-0.0006121455226093531,
... | [
"Conv1d",
"ModelDecoder",
"ModelDecoderBlock",
"Model_hidden_size",
"Module",
"ModuleList",
"__init__",
"block",
"channel_ratios",
"class",
"config",
"conv1",
"conv2",
"conv_blocks",
"def",
"else",
"enumerate",
"for",
"forward",
"hidden_state",
"i",
"if",
"in",
"in_chan... |
xcodec/modeling_xcodec.py:XcodecEuclideanCodebook | [
-0.00022551401343662292,
0.006446313578635454,
0.018369141966104507,
0.00550503795966506,
-0.0009626685059629381,
0.006931213196367025,
0.045181240886449814,
-0.036966472864151,
0.007244972046464682,
0.00036010934854857624,
0.0353691540658474,
0.002681210171431303,
-0.0016115783946588635,
... | [
"F",
"ModelEuclideanCodebook",
"Module",
"Tensor",
"True",
"__init__",
"class",
"clone",
"cluster_size",
"codebook_dim",
"codebook_size",
"config",
"decode",
"def",
"dim",
"dist",
"embed",
"embed_avg",
"embed_ind",
"embedding",
"encode",
"hidden_states",
"indices",
"ini... |
xcodec/modeling_xcodec.py:XcodecVectorQuantization | [
-0.0001221633137902245,
0.018449297174811363,
0.017099348828196526,
0.006159140262752771,
-0.000379673030693084,
0.009505887515842915,
0.0431983545422554,
-0.04139842465519905,
0.002235852414742112,
0.017774323001503944,
0.0431983545422554,
0.009224648587405682,
-0.003346747485920787,
0.00... | [
"ModelEuclideanCodebook",
"ModelVectorQuantization",
"Module",
"__init__",
"class",
"codebook",
"config",
"decode",
"def",
"embed_in",
"embed_ind",
"encode",
"hidden_states",
"nn",
"permute",
"quantize",
"return",
"self",
"super"
] |
xcodec/modeling_xcodec.py:XcodecResidualVectorQuantization | [
-0.00011209374497411773,
0.01124114915728569,
-0.011975495144724846,
-0.003036240115761757,
-0.00028773670783266425,
0.043382927775382996,
0.02632349543273449,
-0.043608881533145905,
0.00610072398558259,
-0.0036717320326715708,
0.029373856261372566,
0.012709842063486576,
0.000476619083201512... | [
"ModelResidualVectorQuantization",
"ModelVectorQuantization",
"Module",
"ModuleList",
"None",
"_",
"__init__",
"all_indices",
"and",
"append",
"bandwidth",
"bw_per_q",
"class",
"codebook_size",
"codes",
"config",
"decode",
"def",
"device",
"embeddings",
"encode",
"enumerate... |
xcodec/modeling_xcodec.py:XcodecPreTrainedModel | [
-0.00004884477675659582,
0.04956846684217453,
-0.010597783140838146,
-0.012111752294003963,
-0.00031190566369332373,
0.01626114919781685,
0.020746983587741852,
-0.043288301676511765,
0.007625918369740248,
-0.016934024170041084,
0.028597192838788033,
0.023887068033218384,
-0.00080955290468409... | [
"AttributeError",
"Conv1d",
"ConvTranspose1d",
"Embedding",
"GroupNorm",
"LayerNorm",
"Linear",
"Model",
"ModelConfig",
"ModelEuclideanCodebook",
"ModelModel",
"ModelPreTrainedModel",
"None",
"PreTrainedAudioTokenizerBase",
"Snake1d",
"Tensor",
"True",
"ValueError",
"__class__",
... |
xcodec/modeling_xcodec.py:XcodecModel | [
-0.00010499545896891505,
0.046529751271009445,
-0.01750512421131134,
-0.008074944838881493,
-0.0003988062671851367,
0.0313962884247303,
0.031848032027482986,
-0.032977394759655,
0.006126793567091227,
-0.022700194269418716,
0.0478849858045578,
0.024394238367676735,
-0.003811599686741829,
-0... | [
"Audio",
"AutoModel",
"ConvTranspose1d",
"F",
"False",
"Identity",
"Linear",
"ModelDecoderOutput",
"ModelEncoderOutput",
"ModelModel",
"ModelOutput",
"ModelPreTrainedModel",
"ModelResidualVectorQuantization",
"None",
"Select",
"SemanticDecoder",
"SemanticEncoder",
"Tanh",
"This",... |
qwen3_vl/modeling_qwen3_vl.py:BaseModelOutputWithDeepstackFeatures | [
-0.000323236221447587,
0.014142042025923729,
0.015544557943940163,
0.02173900604248047,
-0.001380602247081697,
0.05095810070633888,
0.0631132423877716,
-0.030621610581874847,
0.00876572821289301,
-0.011629199609160423,
0.031322870403528214,
0.01601206324994564,
-0.0004565483541227877,
0.01... | [
"ModelModelOutputWithDeepstackFeatures",
"ModelModelOutputWithPooling",
"None",
"class",
"deepstack_features",
"r"
] |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionMLP | [
-0.00020412541925907135,
0.02817285805940628,
0.032262466847896576,
0.033852867782115936,
-0.00048280099872499704,
0.04589449241757393,
0.025673652067780495,
-0.030444862321019173,
0.0004544009279925376,
0.005850411951541901,
0.04589449241757393,
-0.022152045741677284,
0.0011999025009572506,... | [
"ACT2FN",
"Linear",
"ModelVisionMLP",
"Module",
"__init__",
"act_fn",
"class",
"config",
"def",
"forward",
"hidden_act",
"hidden_size",
"hidden_state",
"intermediate_size",
"linear_fc1",
"linear_fc2",
"nn",
"return",
"self",
"super"
] |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionPatchEmbed | [
-0.00006803657743148506,
0.010169493034482002,
0.01101226918399334,
0.00854012742638588,
-0.00019313610391691327,
0.008989606983959675,
0.027081191539764404,
-0.03191310539841652,
0.013877706602215767,
-0.0005021538236178458,
0.008483941666781902,
-0.00022386229829862714,
-0.0015661581419408... | [
"Conv3d",
"ModelVisionPatchEmbed",
"Module",
"__init__",
"class",
"config",
"def",
"dtype",
"embed_dim",
"forward",
"hidden_size",
"hidden_states",
"in_channels",
"kernel_size",
"nn",
"patch_size",
"proj",
"return",
"self",
"stride",
"super",
"target_dtype",
"temporal_pat... |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionRotaryEmbedding | [
-0.00022179921506904066,
0.03434310480952263,
0.01224904041737318,
0.008013390935957432,
-0.0010374479461461306,
0.04464603587985039,
0.0668545737862587,
-0.016713643446564674,
0.005323181394487619,
0.0345720574259758,
0.00887196883559227,
-0.00875749159604311,
-0.0022466115187853575,
0.00... | [
"False",
"ModelVisionRotaryEmbedding",
"Module",
"Tensor",
"__init__",
"arange",
"class",
"def",
"device",
"dim",
"dtype",
"float",
"forward",
"freqs",
"inv_freq",
"nn",
"outer",
"persistent",
"register_buffer",
"return",
"self",
"seq",
"seqlen",
"super",
"theta",
"... |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionPatchMerger | [
-0.00025101262144744396,
0.03235587850213051,
0.011313244700431824,
0.031224554404616356,
-0.0006858654669485986,
0.021042635664343834,
0.00933342706412077,
-0.04683683067560196,
0.00574147142469883,
0.0191193837672472,
0.011200112290680408,
0.007212193217128515,
-0.000643440755084157,
-0.... | [
"False",
"GELU",
"LayerNorm",
"Linear",
"ModelVisionPatchMerger",
"Module",
"__init__",
"act_fn",
"class",
"config",
"def",
"else",
"eps",
"forward",
"hidden_size",
"if",
"linear_fc1",
"linear_fc2",
"nn",
"norm",
"out_hidden_size",
"return",
"self",
"spatial_merge_size"... |
qwen3_vl/modeling_qwen3_vl.py:rotate_half | [
0.00002049485374300275,
0.013860220089554787,
0.03434192016720772,
0.002974055241793394,
0.0003507140791043639,
0.028506038710474968,
0.01975221559405327,
-0.02008890174329281,
0.014477476477622986,
0.018742159008979797,
-0.001487027620896697,
-0.01217679213732481,
0.00022445700597018003,
... | [
"Model_half",
"cat",
"def",
"dim",
"return",
"shape",
"torch",
"x",
"x1",
"x2"
] |
qwen3_vl/modeling_qwen3_vl.py:apply_rotary_pos_emb_vision | [
-0.00006463831959990785,
0.03513491153717041,
0.018587501719594002,
-0.0022809358779340982,
-0.0004639792023226619,
0.01700076460838318,
0.060749396681785583,
-0.009180412627756596,
0.016887426376342773,
0.04692211002111435,
0.009010405279695988,
0.01983422413468361,
0.0013458938337862492,
... | [
"Model_rotary_pos_emb_vision",
"cos",
"def",
"dtype",
"float",
"k",
"k_embed",
"orig_q_dtype",
"q",
"q_embed",
"return",
"rotate_half",
"sin",
"unsqueeze"
] |
qwen3_vl/modeling_qwen3_vl.py:repeat_kv | [
-0.00025096136960200965,
-0.0023088448215276003,
-0.004072744864970446,
-0.009292741306126118,
-0.000559285341296345,
0.03143470734357834,
0.009694280102849007,
-0.058739304542541504,
0.011185707524418831,
0.05162634328007698,
0.005736260209232569,
-0.02317449077963829,
0.0006775957299396396... | [
"Model_kv",
"None",
"batch",
"def",
"expand",
"head_dim",
"hidden_states",
"if",
"n_rep",
"num_key_value_heads",
"reshape",
"return",
"shape",
"slen"
] |
qwen3_vl/modeling_qwen3_vl.py:eager_attention_forward | [
0,
0.021594731137156487,
0.01775064319372177,
-0.018768195062875748,
-0.00008832923776935786,
0.039797618985176086,
0.05359111353754997,
-0.035049039870500565,
0.02069023996591568,
0.011645326390862465,
0.029395969584584236,
0.022499222308397293,
0.002741739386692643,
-0.014584923163056374... | [
"Model_attention_forward",
"None",
"attention_mask",
"attn_output",
"attn_weights",
"causal_mask",
"contiguous",
"def",
"dim",
"dropout",
"dtype",
"float32",
"functional",
"if",
"is",
"key",
"key_states",
"kwargs",
"matmul",
"module",
"nn",
"not",
"num_key_value_groups",
... |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionAttention | [
-0.00014312515850178897,
0.03664004057645798,
0.021825702860951424,
0.021825702860951424,
-0.0004081717343069613,
0.0252182986587286,
0.03958028927445412,
-0.029628673568367958,
0.0055129691027104855,
0.03189040347933769,
0.028837068006396294,
0.03189040347933769,
0.0010036430321633816,
-0... | [
"ALL_ATTENTION_FUNCTIONS",
"False",
"Linear",
"ModelVisionAttention",
"Module",
"None",
"Tensor",
"_",
"__init__",
"_attn_implementation",
"apply_rotary_pos_emb_vision",
"attention_dropout",
"attention_interface",
"attention_mask",
"attn_output",
"attn_outputs",
"cat",
"class",
"... |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionBlock | [
-0.00016528498963452876,
0.013067236170172691,
0.01606534607708454,
0.01855434477329254,
-0.00040481562609784305,
0.043896861374378204,
0.026587016880512238,
-0.01866747997701168,
0.005232551135122776,
0.021043341606855392,
0.021043341606855392,
0.019120024517178535,
0.0016475465381518006,
... | [
"GradientCheckpointingLayer",
"LayerNorm",
"ModelVisionAttention",
"ModelVisionBlock",
"ModelVisionMLP",
"None",
"Tensor",
"__init__",
"attn",
"attn_implementation",
"class",
"config",
"cu_seqlens",
"def",
"eps",
"forward",
"hidden_size",
"hidden_states",
"kwargs",
"mlp",
"nn... |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextRotaryEmbedding | [
-0.00028448071680031717,
0.04476313292980194,
0.002913662465289235,
-0.02087399922311306,
-0.0013263686560094357,
0.05728752911090851,
0.032470665872097015,
-0.017163066193461418,
-0.007769766263663769,
0.026440398767590523,
0.02052609995007515,
0.0017250040546059608,
-0.0010509478161111474,... | [
"False",
"ModelTextRotaryEmbedding",
"Module",
"None",
"ROPE_INIT_FUNCTIONS",
"Tensor",
"__init__",
"and",
"apply_interleaved_mrope",
"arange",
"attention_factor",
"attention_scaling",
"base",
"cat",
"class",
"clone",
"compute_default_rope_parameters",
"config",
"cos",
"cpu",
... |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextRMSNorm | [
-0.0000997777096927166,
0.04746946692466736,
0.03865370899438858,
0.048825737088918686,
-0.0005686446093022823,
0.03435885161161423,
0.027125408872961998,
-0.03481094166636467,
0.008702735416591167,
0.035715121775865555,
0.021474281325936317,
0.011415275745093822,
0.002090916968882084,
0.0... | [
"ModelTextRMSNorm",
"Module",
"Parameter",
"True",
"__init__",
"class",
"def",
"dtype",
"eps",
"extra_repr",
"f",
"float32",
"forward",
"hidden_size",
"hidden_states",
"input_dtype",
"keepdim",
"mean",
"nn",
"ones",
"pow",
"return",
"rsqrt",
"self",
"shape",
"super"... |
qwen3_vl/modeling_qwen3_vl.py:apply_rotary_pos_emb | [
-0.0001444592053303495,
0.027112245559692383,
0.028019767254590988,
0.0028360087890177965,
-0.0005707467789761722,
0.021667107939720154,
0.046510547399520874,
-0.002183726755902171,
0.01293220091611147,
0.03652779385447502,
0.007884104736149311,
0.0017654155381023884,
-0.0007834474672563374,... | [
"Model_rotary_pos_emb",
"cos",
"def",
"k",
"k_embed",
"q",
"q_embed",
"return",
"rotate_half",
"sin",
"unsqueeze",
"unsqueeze_dim"
] |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextAttention | [
-0.00005613891698885709,
0.046707578003406525,
0.0357043519616127,
0.0008385750697925687,
-0.00038420071359723806,
0.025936178863048553,
0.033234238624572754,
-0.027171235531568527,
0.0009052400127984583,
0.01998545415699482,
0.022455567494034767,
0.037276241928339005,
-0.0011087435996159911... | [
"ALL_ATTENTION_FUNCTIONS",
"Linear",
"ModelTextAttention",
"ModelTextRMSNorm",
"Module",
"None",
"Tensor",
"True",
"__init__",
"_attn_implementation",
"apply_rotary_pos_emb",
"attention_dropout",
"attention_interface",
"attention_mask",
"attn_output",
"attn_weights",
"cache_kwargs",
... |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextMLP | [
-0.00019116139446850866,
0.03247242420911789,
0.03338714316487312,
0.01726526953279972,
-0.0005788438720628619,
0.057398442178964615,
0.02641243115067482,
-0.016922250390052795,
-0.002029526513069868,
-0.008232446387410164,
0.035216573625802994,
-0.026069412007927895,
-0.0016507769469171762,... | [
"ACT2FN",
"Linear",
"ModelTextMLP",
"Module",
"__init__",
"act_fn",
"class",
"config",
"def",
"down_proj",
"forward",
"gate_proj",
"hidden_act",
"hidden_size",
"intermediate_size",
"nn",
"return",
"self",
"super",
"up_proj",
"x"
] |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextDecoderLayer | [
-0.00016621610848233104,
0.0436830073595047,
0.024897051975131035,
0.0035789511166512966,
-0.0007568137953057885,
0.039382610470056534,
0.036666568368673325,
-0.03825092315673828,
0.005573544651269913,
-0.0032252997625619173,
0.01007198914885521,
0.028518440201878548,
-0.002178492024540901,
... | [
"False",
"GradientCheckpointingLayer",
"ModelTextAttention",
"ModelTextDecoderLayer",
"ModelTextMLP",
"ModelTextRMSNorm",
"None",
"Tensor",
"_",
"__init__",
"attention_mask",
"cache_position",
"class",
"config",
"def",
"eps",
"forward",
"hidden_size",
"hidden_states",
"input_la... |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLModelOutputWithPast | [
-0.0002648343506734818,
0.008071142248809338,
0.014355102553963661,
0.0052174171432852745,
-0.0014196563279256225,
0.042200542986392975,
0.04911866411566734,
-0.03851087763905525,
0.016949398443102837,
0.0025078190956264734,
0.008071142248809338,
0.03182335942983627,
-0.001830419758334756,
... | [
"ModelModelOutputWithPast",
"ModelOutput",
"None",
"attentions",
"class",
"hidden_states",
"last_hidden_state",
"past_key_values",
"r",
"rope_deltas"
] |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLPreTrainedModel | [
-0.0003329527680762112,
0.04890584945678711,
0.004803253337740898,
0.010014055296778679,
-0.0017757480964064598,
0.032836783677339554,
0.02724754437804222,
-0.034234095364809036,
-0.0043083722703158855,
0.015719737857580185,
0.009431842714548111,
-0.0039881556294858456,
-0.005560129415243864... | [
"ModelConfig",
"ModelPreTrainedModel",
"ModelTextAttention",
"ModelTextDecoderLayer",
"ModelVisionBlock",
"ModelVisionRotaryEmbedding",
"PreTrainedModel",
"True",
"_can_compile_fullgraph",
"_can_record_outputs",
"_init_weights",
"_no_split_modules",
"_skip_keys_device_placement",
"_support... |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionModel | [
-0.0002279315231135115,
0.01696234568953514,
-0.023521119728684425,
0.01933707483112812,
-0.000816312909591943,
0.03460318595170975,
0.047494567930698395,
-0.05156553164124489,
0.0019223992712795734,
0.02159872092306614,
0.01809316873550415,
0.00689802085980773,
-0.0006749600288458169,
0.0... | [
"BaseModelOutputWithDeepstackFeatures",
"Embedding",
"F",
"False",
"ModelPreTrainedModel",
"ModelVisionAttention",
"ModelVisionBlock",
"ModelVisionConfig",
"ModelVisionModel",
"ModelVisionPatchEmbed",
"ModelVisionPatchMerger",
"ModelVisionRotaryEmbedding",
"ModuleList",
"None",
"T",
"T... |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextModel | [
-0.00009935703565133736,
0.04078628122806549,
0.0032774689607322216,
0.002044916618615389,
-0.0006582950591109693,
0.03159816190600395,
0.03854527696967125,
-0.04190678521990776,
0.004874184727668762,
0.008291716687381268,
0.03159816190600395,
0.020505191758275032,
-0.002212991937994957,
0... | [
"BaseModelOutputWithPast",
"DynamicCache",
"Embedding",
"False",
"ModelPreTrainedModel",
"ModelTextConfig",
"ModelTextDecoderLayer",
"ModelTextModel",
"ModelTextRMSNorm",
"ModelTextRotaryEmbedding",
"ModuleList",
"None",
"ValueError",
"You",
"__init__",
"_deepstack_process",
"_no_spl... |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLModel | [
-0.00008527900354238227,
0.03896069899201393,
-0.018472744151949883,
0.0014134448720142245,
-0.0003113777202088386,
0.05396280437707901,
0.03269116207957268,
-0.0470215305685997,
0.005877691321074963,
0.022055337205529213,
0.033586807548999786,
0.020264040678739548,
0.0004110885492991656,
... | [
"False",
"Image",
"ModelConfig",
"ModelModel",
"ModelModelOutputWithPast",
"ModelPreTrainedModel",
"ModelTextDecoderLayer",
"ModelTextModel",
"ModelVisionBlock",
"ModelVisionModel",
"None",
"True",
"ValueError",
"Video",
"You",
"_",
"__init__",
"_checkpoint_conversion_mapping",
"... |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLCausalLMOutputWithPast | [
-0.0002988479973282665,
0.017503436654806137,
0.020285440608859062,
-0.0002644352207425982,
-0.0015503872418776155,
0.049844224005937576,
0.049612391740083694,
-0.02874736674129963,
0.016923854127526283,
0.0007607040461152792,
0.017155686393380165,
0.02156052552163601,
-0.0010577408829703927... | [
"ModelCausalLMOutputWithPast",
"ModelOutput",
"None",
"attentions",
"class",
"hidden_states",
"logits",
"loss",
"past_key_values",
"r",
"rope_deltas"
] |
qwen3_vl/modeling_qwen3_vl.py:Qwen3VLForConditionalGeneration | [
-0.00028949574334546924,
0.03455477952957153,
-0.010002699680626392,
0.019778065383434296,
-0.000994586618617177,
0.03910146281123161,
0.031826771795749664,
-0.050922833383083344,
-0.0006997627206146717,
0.01182137243449688,
0.03523678332567215,
0.00115087884478271,
0.0005896477960050106,
... | [
"False",
"GenerationMixin",
"Linear",
"ModelCausalLMOutputWithPast",
"ModelConfig",
"ModelForConditionalGeneration",
"ModelModel",
"ModelPreTrainedModel",
"None",
"Tensor",
"True",
"__init__",
"_checkpoint_conversion_mapping",
"_expand_dict_for_generation",
"_expand_dict_for_generation_v... |
falcon_mamba/modeling_falcon_mamba.py:FalconMambaCache | [
-0.00018014491070061922,
0.024301903322339058,
0.0095512131229043,
-0.014581141993403435,
-0.0004415316507220268,
0.006697152275592089,
-0.013733400963246822,
-0.05222083255648613,
0.006188507657498121,
0.031648989766836166,
0.012998691760003567,
-0.0025573514867573977,
0.0011656435672193766... | [
"ModelCache",
"None",
"True",
"_",
"__init__",
"_dtype",
"_dynamo",
"append",
"cache_position",
"clamp",
"class",
"config",
"conv_kernel",
"conv_kernel_size",
"conv_state",
"conv_states",
"def",
"device",
"dims",
"dtype",
"else",
"float16",
"for",
"if",
"in",
"inter... |
falcon_mamba/modeling_falcon_mamba.py:rms_forward | [
-0.00017206810298375785,
0.03139933943748474,
0.039757438004016876,
0.03907975181937218,
-0.00045884825522080064,
0.007510993164032698,
0.029592182487249374,
-0.032980602234601974,
0.01248067244887352,
0.032980602234601974,
0.03456186130642891,
-0.0034166546538472176,
0.0024142477195709944,
... | [
"Model_forward",
"True",
"def",
"float32",
"hidden_states",
"keepdim",
"mean",
"pow",
"return",
"rsqrt",
"to",
"torch",
"variance",
"variance_epsilon"
] |
falcon_mamba/modeling_falcon_mamba.py:FalconMambaMixer | [
-0.0001887798571260646,
0.022454118356108665,
0.00946036446839571,
0.0008370428113266826,
-0.00034550277632661164,
0.05015132948756218,
0.013905595988035202,
-0.035333890467882156,
0.008662502281367779,
0.010144246742129326,
0.020174512639641762,
-0.0016028479440137744,
0.001937665045261383,... | [
"A",
"ACT2FN",
"A_log",
"B",
"C",
"Conv1d",
"D",
"Falling",
"False",
"ImportError",
"Linear",
"Mamba",
"Model",
"ModelMixer",
"Model_inner_fn",
"Model_ssm",
"Module",
"None",
"Parameter",
"The",
"To",
"True",
"_",
"__init__",
"_is_quantized",
"act",
"activation",
... |
falcon_mamba/modeling_falcon_mamba.py:FalconMambaRMSNorm | [
-0.00008794592577032745,
0.03575112298130989,
0.02059083804488182,
0.05475805327296257,
-0.00028284118161536753,
0.044802043586969376,
0.02138279192149639,
-0.028962936252355576,
0.008654939942061901,
0.0366562157869339,
0.014481468126177788,
0.0026304228231310844,
0.0032385315280407667,
0... | [
"ModelRMSNorm",
"Module",
"Parameter",
"__init__",
"class",
"def",
"device",
"eps",
"extra_repr",
"f",
"forward",
"hidden_size",
"hidden_states",
"nn",
"ones",
"return",
"rms_forward",
"self",
"shape",
"super",
"to",
"torch",
"variance_epsilon",
"weight"
] |
falcon_mamba/modeling_falcon_mamba.py:FalconMambaBlock | [
-0.00018491271475795656,
-0.0019252155907452106,
0.02095087617635727,
0.004699791315943003,
-0.0003945984353777021,
0.05096159130334854,
0.02672652341425419,
-0.03623935207724571,
0.004190175328403711,
0.015401724725961685,
-0.005350966937839985,
-0.007587614469230175,
0.000422910408815369,
... | [
"GradientCheckpointingLayer",
"ModelBlock",
"ModelMixer",
"ModelRMSNorm",
"None",
"__init__",
"attention_mask",
"cache_params",
"cache_position",
"class",
"config",
"def",
"dtype",
"eps",
"float32",
"forward",
"hidden_size",
"hidden_states",
"if",
"layer_idx",
"layer_norm_eps... |
falcon_mamba/modeling_falcon_mamba.py:FalconMambaPreTrainedModel | [
-0.00028834445402026176,
0.028934119269251823,
-0.004471118561923504,
0.0049837310798466206,
-0.0011818562634289265,
0.0323515348136425,
0.011505299247801304,
-0.003146870294585824,
-0.009170065633952618,
0.020846234634518623,
-0.00032750231912359595,
-0.03349067270755768,
-0.002748171798884... | [
"A",
"A_log",
"D",
"Embedding",
"Linear",
"ModelBlock",
"ModelConfig",
"ModelMixer",
"ModelPreTrainedModel",
"ModelRMSNorm",
"None",
"PreTrainedModel",
"True",
"_init_weights",
"_is_stateful",
"_no_split_modules",
"a",
"arange",
"b_c_rms",
"backbone",
"base_model_prefix",
"... |
falcon_mamba/modeling_falcon_mamba.py:FalconMambaOutput | [
-0.00017384765669703484,
0.02076900005340576,
0.031952306628227234,
0.02179603837430477,
-0.001112625002861023,
0.03309345990419388,
0.05135192349553108,
-0.01997019164264202,
0.017916115000844002,
-0.016204385086894035,
-0.003166701877489686,
0.007988076657056808,
-0.0008736958843655884,
... | [
"ModelOutput",
"None",
"cache_params",
"class",
"hidden_states",
"last_hidden_state",
"r"
] |
falcon_mamba/modeling_falcon_mamba.py:FalconMambaCausalLMOutput | [
-0.0001988249714486301,
0.028416814282536507,
0.035378362983465195,
0.004079923965036869,
-0.0011483702110126615,
0.03857382759451866,
0.050899192690849304,
-0.007817476987838745,
0.01791742816567421,
-0.015977324917912483,
0.0023538023233413696,
-0.0003940835886169225,
-0.002425129758194089... | [
"ModelCausalLMOutput",
"ModelOutput",
"None",
"cache_params",
"class",
"hidden_states",
"logits",
"loss",
"r"
] |
falcon_mamba/modeling_falcon_mamba.py:FalconMambaModel | [
0.000011523719877004623,
0.0359627865254879,
-0.0007340061129070818,
0.00797924306243658,
0.00009877474076347426,
0.040682900696992874,
0.027758775278925896,
0.0007550780428573489,
0.005310130305588245,
0.0076420921832323074,
0.012980318628251553,
-0.013879387639462948,
0.0006497183348983526... | [
"Embedding",
"False",
"ModelBlock",
"ModelCache",
"ModelModel",
"ModelOutput",
"ModelPreTrainedModel",
"ModelRMSNorm",
"ModuleList",
"None",
"True",
"ValueError",
"You",
"__init__",
"a",
"all_hidden_states",
"and",
"arange",
"are",
"attention_mask",
"auto_docstring",
"autom... |
falcon_mamba/modeling_falcon_mamba.py:FalconMambaForCausalLM | [
-0.00022307969629764557,
0.030140545219182968,
0.014617031440138817,
-0.00611875718459487,
-0.0008994006784632802,
0.019602686166763306,
0.012747411616146564,
-0.007648446597158909,
-0.0009418920380994678,
0.007875067181885242,
0.012917377054691315,
-0.013200652785599232,
0.00167840917129069... | [
"Any",
"CrossEntropyLoss",
"False",
"GenerationMixin",
"Linear",
"ModelCache",
"ModelCausalLMOutput",
"ModelForCausalLM",
"ModelModel",
"ModelPreTrainedModel",
"Model_outputs",
"None",
"True",
"__init__",
"_tied_weights_keys",
"_update_model_kwargs_for_generation",
"and",
"arange",... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.