identifier
stringlengths
24
102
embedding
listlengths
2.56k
2.56k
tokens
listlengths
4
448
tapas/modeling_tapas.py:reduce_mean
[ -0.0000661029334878549, -0.005499763879925013, -0.00023444507678505033, 0.05144394561648369, -0.0001877323375083506, 0.04196742922067642, 0.02978333830833435, 0.014383998699486256, 0.006176657974720001, 0.003595999674871564, 0.017824877053499222, 0.003046023193746805, -0.0030319211073219776,...
[ "Model_mean", "_segment_Model", "def", "index", "mean", "name", "return", "segmented_Model_mean", "values" ]
tapas/modeling_tapas.py:reduce_max
[ -0.00004356446515885182, 0.0008378459024243057, 0.025571901351213455, 0.04573652893304825, -0.0002798686909954995, 0.028275538235902786, 0.025459250435233116, 0.011377806775271893, 0.00833621434867382, 0.002619148464873433, 0.030415918678045273, -0.03447137400507927, -0.002126297913491726, ...
[ "Model_max", "_segment_Model", "amax", "def", "index", "name", "return", "segmented_Model_max", "values" ]
tapas/modeling_tapas.py:reduce_min
[ 0.000237918458878994, 0.01568472944200039, 0.02312638983130455, 0.03755176067352295, 0.0009230520226992667, 0.04968738928437233, 0.03984150290489197, -0.004236022010445595, 0.015226781368255615, 0.021638058125972748, 0.012536334805190563, -0.017974471673369408, 0.0006082126055844128, 0.009...
[ "Model_min", "_segment_Model", "amin", "def", "index", "name", "return", "segmented_Model_min", "values" ]
tapas/modeling_tapas.py:compute_column_logits
[ -0.0002476426598150283, 0.0228045042604208, -0.004475384019315243, 0.01328362338244915, -0.0008480424876324832, 0.032610442489385605, 0.016875332221388817, -0.004874462727457285, 0.002479989780113101, 0.0139107471331954, 0.023146571591496468, 0.02451484277844429, -0.0057011260651052, -0.00...
[ "CLOSE_ENOUGH_TO_LOG_ZERO", "EPSILON_ZERO_DIVISION", "Model_column_logits", "_", "allow_empty_column_selection", "as_tensor", "bs", "bsj", "cell_count", "cell_index", "cell_logits", "cell_logits_index", "cell_mask", "column_index", "column_logits", "column_output_bias", "column_outpu...
tapas/modeling_tapas.py:_single_column_cell_selection_loss
[ -0.00026278203586116433, 0.03732224553823471, -0.0007955455803312361, -0.01635728031396866, -0.0011735197622328997, 0.03133225813508034, 0.008293832652270794, 0.00570200989022851, -0.0012527143117040396, 0.0035277586430311203, 0.024881497025489807, 0.008639409206807613, -0.005097251385450363...
[ "Bernoulli", "CLOSE_ENOUGH_TO_LOG_ZERO", "Categorical", "EPSILON_ZERO_DIVISION", "_", "_single_column_cell_selection_loss", "argmax", "as_tensor", "cell_dist", "cell_index", "cell_log_prob", "cell_loss", "cell_mask", "col_index", "column_dist", "column_id_for_cells", "column_label", ...
tapas/modeling_tapas.py:compute_token_logits
[ -0.0001195065415231511, 0.047590162605047226, 0.024588249623775482, -0.021868811920285225, -0.00046386243775486946, 0.040564946830272675, 0.06753270328044891, -0.0035409345291554928, 0.010877750813961029, 0.006118734832853079, 0.03353973478078842, 0.015070217661559582, -0.0006409091874957085...
[ "Model_token_logits", "bs", "bsj", "def", "einsum", "j", "logits", "output_bias", "output_weights", "return", "sequence_output", "temperature", "torch" ]
tapas/modeling_tapas.py:_calculate_aggregate_mask
[ -0.0003667731652967632, 0.013886540196835995, 0.003892879467457533, -0.0019173884065821767, -0.0016777148703113198, 0.049968305975198746, -0.01940629445016384, -0.014409464783966541, 0.0034425838384777308, -0.0028179800137877464, 0.04183392971754074, -0.0009732198668643832, -0.00203359383158...
[ "Categorical", "FloatTensor", "_calculate_aggregate_mask", "aggregate_mask", "aggregate_mask_init", "aggregation_classifier", "aggregation_ops_total_mass", "answer", "categorical", "cell_selection_preference", "def", "detach", "device", "dim", "dist_aggregation", "distributions", "dt...
tapas/modeling_tapas.py:_calculate_aggregation_loss_known
[ -0.0003723987319972366, 0.011164731346070766, 0.00954497791826725, 0.0024440926499664783, -0.0014462086837738752, 0.05645998939871788, -0.013594361953437328, 0.005003882106393576, 0.0063633183017373085, -0.014172845520079136, 0.04188220575451851, 0.01255309209227562, 0.0013811293756589293, ...
[ "_calculate_aggregation_loss_known", "aggregate_mask", "aggregation_labels", "def", "dim", "dtype", "else", "float32", "functional", "if", "log_probs", "log_softmax", "logits_aggregation", "long", "nn", "num_aggregation_labels", "num_classes", "one_hot", "one_hot_labels", "per_...
tapas/modeling_tapas.py:_calculate_aggregation_loss_unknown
[ -0.00014795942115597427, 0.01657145470380783, 0.012344042770564556, -0.017022378742694855, -0.0006834315718151629, 0.05117986723780632, -0.022997120395302773, -0.004481056705117226, 0.010765808634459972, -0.016458723694086075, 0.03449568152427673, 0.012174946255981922, -0.0003364315198268741...
[ "Categorical", "_calculate_aggregation_loss_unknown", "aggregate_mask", "aggregation_ops_total_mass", "categorical", "def", "dim", "dist_aggregation", "distributions", "log", "logits", "logits_aggregation", "probs", "return", "sum", "torch" ]
tapas/modeling_tapas.py:_calculate_aggregation_loss
[ -0.000329378031892702, 0.006080825347453356, 0.008976456709206104, -0.0007782008615322411, -0.0016070753335952759, 0.06347223371267319, -0.012277476489543915, -0.003851189510896802, 0.008223592303693295, -0.029187962412834167, 0.04609844833612442, 0.017257962375879288, 0.004517184570431709, ...
[ "_calculate_aggregation_loss", "_calculate_aggregation_loss_known", "_calculate_aggregation_loss_unknown", "aggregate_mask", "aggregation_labels", "aggregation_loss_weight", "def", "if", "logits_aggregation", "num_aggregation_labels", "per_example_aggregation_loss", "return", "use_answer_as_...
tapas/modeling_tapas.py:_calculate_expected_result
[ -0.00048544409219175577, 0.04980434104800224, -0.010731649585068226, 0.019091663882136345, -0.0020751808770000935, 0.03177991136908531, 0.025257915258407593, 0.011739594861865044, -0.006166251841932535, 0.009960868395864964, 0.033440057188272476, -0.0355745293200016, -0.002475394168868661, ...
[ "AverageApproximationFunction", "EPSILON_ZERO_DIVISION", "FIRST_ORDER", "Invalid", "RATIO", "RelaxedBernoulli", "RelaxedOneHotCategorical", "SECOND_ORDER", "True", "ValueError", "_calculate_expected_result", "aggregation_op_only_probs", "aggregation_temperature", "all_results", "average_...
tapas/modeling_tapas.py:huber_loss
[ -0.00010970180301228538, 0.02853480540215969, 0.043760884553194046, -0.020978156477212906, -0.00017094095528591424, 0.06225775554776192, 0.016692297533154488, 0.006541575770825148, 0.012857579626142979, 0.03022659197449684, 0.030903305858373642, -0.027294160798192024, 0.0021852245554327965, ...
[ "Model_loss", "abs", "def", "delta", "errors", "input", "return", "target", "torch", "where" ]
tapas/modeling_tapas.py:_calculate_regression_loss
[ -0.0003564467479009181, 0.030877649784088135, 0.0398644283413887, -0.0062216161750257015, -0.001360978465527296, 0.050925079733133316, 0.0055303252302110195, 0.028227701783180237, -0.000986529397778213, 0.018088772892951965, 0.033873241394758224, 0.016936620697379112, 0.00045005904394201934,...
[ "EPSILON_ZERO_DIVISION", "None", "_calculate_expected_result", "_calculate_regression_loss", "abs", "aggregate_mask", "answer", "answer_loss_cutoff", "answer_loss_importance", "answer_masked", "config", "def", "delta", "detach", "dist_per_cell", "dtype", "else", "expected_result", ...
smolvlm/modeling_smolvlm.py:SmolVLMPreTrainedModel
[ -0.0003620385832618922, 0.034697484225034714, 0.016650136560201645, 0.01874595694243908, -0.0018411207711324096, 0.03027297370135784, 0.02549915947020054, -0.029690802097320557, -0.002052158350124955, 0.01513648685067892, 0.013739272952079773, 0.010770192369818687, -0.0027362110558897257, ...
[ "ModelConfig", "ModelDecoderLayer", "ModelPreTrainedModel", "ModelVisionAttention", "PreTrainedModel", "True", "_no_split_modules", "_skip_keys_device_placement", "_supports_attention_backend", "_supports_flash_attn", "_supports_flex_attn", "_supports_sdpa", "base_model_prefix", "class", ...
smolvlm/modeling_smolvlm.py:SmolVLMVisionEmbeddings
[ -0.00005818148565595038, 0.008287143893539906, 0.013998553156852722, 0.012822674587368965, -0.00019947938562836498, 0.012934663332998753, 0.02911699004471302, -0.006103369407355785, 0.01545440312474966, 0.009463022463023663, 0.03337255120277405, 0.03874799609184265, 0.0017358206678181887, ...
[ "Conv2d", "Embedding", "ModelVisionEmbeddings", "Module", "None", "True", "_", "__init__", "arange", "batch_size", "boundaries", "bucket_coords_h", "bucket_coords_w", "bucketize", "clamp", "class", "config", "def", "device", "dim", "dtype", "embed_dim", "embeddings", "f...
smolvlm/modeling_smolvlm.py:eager_attention_forward
[ 0.000051929029723396525, 0.031000085175037384, 0.02477744035422802, -0.010239078663289547, 0.00008883178816176951, 0.035978201776742935, 0.05928483232855797, -0.020365018397569656, 0.022175243124365807, 0.01170988567173481, 0.02477744035422802, 0.031452640891075134, 0.0025173425674438477, ...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "softmax", "...
smolvlm/modeling_smolvlm.py:SmolVLMVisionAttention
[ -0.00013196600775700063, 0.03941384702920914, 0.03626073896884918, 0.013851151801645756, -0.00018827150051947683, 0.019594311714172363, 0.04459395259618759, -0.01869342476129532, 0.0008868115255609155, 0.043017398566007614, 0.024661807343363762, 0.013062874786555767, -0.0029138093814253807, ...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "ModelVisionAttention", "Module", "None", "ValueError", "__init__", "_attn_implementation", "and", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "batch_size", "be", "by", "class", "conf...
smolvlm/modeling_smolvlm.py:SmolVLMVisionMLP
[ -0.0001662031572777778, 0.03708730265498161, 0.032536715269088745, 0.034129418432712555, -0.00047283468302339315, 0.04709859937429428, 0.03162659704685211, -0.03481200709939003, 0.003341838950291276, 0, 0.049373894929885864, -0.020136358216404915, 0.0019482210045680404, -0.0018913387320935...
[ "ACT2FN", "Linear", "ModelVisionMLP", "Module", "__init__", "activation_fn", "class", "config", "def", "fc1", "fc2", "forward", "hidden_act", "hidden_size", "hidden_states", "intermediate_size", "nn", "return", "self", "super" ]
smolvlm/modeling_smolvlm.py:SmolVLMEncoderLayer
[ -0.00010839982860488817, 0.028088776394724846, 0.033616285771131516, 0.024591779336333275, -0.0000925364438444376, 0.0489579476416111, 0.03090893290936947, -0.027750356122851372, 0.011111423373222351, 0.021884428337216377, 0.012183083221316338, 0.027524743229150772, 0.00019917366444133222, ...
[ "GradientCheckpointingLayer", "LayerNorm", "ModelEncoderLayer", "ModelVisionAttention", "ModelVisionMLP", "_", "__init__", "attention_mask", "auto_docstring", "class", "config", "def", "embed_dim", "eps", "forward", "hidden_size", "hidden_states", "kwargs", "layer_norm1", "laye...
smolvlm/modeling_smolvlm.py:SmolVLMEncoder
[ -0.00006501419557025656, 0.016193805262446404, 0.02575264871120453, 0.03801045939326286, -0.00009224986570188776, 0.03958485648036003, 0.009952442720532417, -0.019792428240180016, 0.007815760560333729, 0.005510392133146524, 0.009109015576541424, 0.003851651679724455, -0.002038282807916403, ...
[ "BaseModelOutput", "False", "ModelEncoder", "ModelEncoderLayer", "Module", "ModuleList", "None", "_", "__init__", "attention_mask", "auto_docstring", "class", "config", "def", "encoder_layer", "for", "forward", "gradient_checkpointing", "hidden_states", "in", "inputs_embeds",...
smolvlm/modeling_smolvlm.py:SmolVLMVisionTransformer
[ -0.00003304134952486493, 0.04884692654013634, 0.028680766001343727, 0.019830061122775078, 0, 0.016356999054551125, 0.03831570968031883, -0.023975327610969543, 0.0026608132757246494, 0.013668176718056202, 0.012435800395905972, 0.012883937917649746, -0.0017155242385342717, -0.010699269361793...
[ "BaseModelOutput", "False", "LayerNorm", "ModelEncoder", "ModelEncoderLayer", "ModelPreTrainedModel", "ModelVisionAttention", "ModelVisionConfig", "ModelVisionEmbeddings", "ModelVisionTransformer", "None", "True", "__init__", "_can_record_outputs", "_supports_flash_attn", "_supports_fl...
smolvlm/modeling_smolvlm.py:SmolVLMBaseModelOutputWithPast
[ -0.00023066798166837543, 0.013790011405944824, 0.01911146752536297, 0.021514706313610077, -0.001316059147939086, 0.027007821947336197, 0.040282852947711945, -0.04440269246697426, 0.015907149761915207, -0.006065316032618284, 0.009155194275081158, 0.026893382892012596, -0.0026607282925397158, ...
[ "ModelBaseModelOutputWithPast", "ModelOutput", "None", "attentions", "class", "hidden_states", "image_hidden_states", "last_hidden_state", "past_key_values", "r" ]
smolvlm/modeling_smolvlm.py:SmolVLMSimpleMLP
[ -0.0001559507945785299, 0.02404477633535862, 0.034932978451251984, 0.05194579064846039, -0.00037924398202449083, 0.04468699172139168, 0.02903520129621029, -0.03425246477127075, 0, 0.014234053902328014, 0.03538665175437927, -0.024271614849567413, -0.001999005675315857, 0.0012688722927123308...
[ "Linear", "ModelSimpleMLP", "Module", "__init__", "class", "config", "def", "forward", "hidden_size", "input_size", "nn", "output_size", "proj", "return", "scale_factor", "self", "super", "text_config", "vision_config", "x" ]
smolvlm/modeling_smolvlm.py:SmolVLMConnector
[ -0.00030849099857732654, 0.015292594209313393, 0.020998787134885788, 0.015292594209313393, -0.0011127074249088764, 0.03651962801814079, 0.024650748819112778, -0.06117037683725357, 0.00309560913592577, 0.023509511724114418, 0.018488062545657158, -0.013295426964759827, -0.0037090247496962547, ...
[ "ModelConnector", "ModelSimpleMLP", "Module", "__init__", "bsz", "class", "config", "def", "embed_dim", "forward", "height", "image_hidden_states", "int", "modality_projection", "nn", "permute", "pixel_shuffle", "reshape", "return", "scale_factor", "self", "seq", "size", ...
smolvlm/modeling_smolvlm.py:SmolVLMModel
[ 0.000011549662303877994, 0.04286540672183037, 0.006996167358011007, 0.009883474558591843, 0.00019780830189120024, 0.030205674469470978, 0.01465863548219204, -0.031094076111912727, 0.003720184089615941, 0.01621333882212639, 0.033759281039237976, 0.02609681338071823, 0.0011452059261500835, 0...
[ "At", "AutoModel", "DynamicCache", "Encodes", "False", "ModelBaseModelOutputWithPast", "ModelConnector", "ModelModel", "ModelPreTrainedModel", "ModelVisionTransformer", "None", "Setting", "True", "_", "__init__", "_from_config", "all", "and", "any", "attention_mask", "attenti...
smolvlm/modeling_smolvlm.py:SmolVLMCausalLMOutputWithPast
[ -0.00020158913685008883, 0.02637176588177681, 0.02978190779685974, 0.009434727020561695, -0.0011793408775702119, 0.03387407958507538, 0.03773890808224678, -0.030463937669992447, 0.01784641109406948, -0.004205842036753893, 0.020006166771054268, 0.018642110750079155, -0.0008596400148235261, ...
[ "ModelCausalLMOutputWithPast", "ModelOutput", "None", "attentions", "class", "hidden_states", "image_hidden_states", "logits", "loss", "past_key_values", "r" ]
smolvlm/modeling_smolvlm.py:SmolVLMForConditionalGeneration
[ -0.00017000327352434397, 0.04127749428153038, 0.01592773012816906, 0.013403969816863537, -0.0005818668869324028, 0.02063874714076519, 0.02983645163476467, -0.03948282077908516, 0.000683518301229924, 0.011609295383095741, 0.04217483103275299, 0.012618799693882465, 0.0011567232431843877, 0.0...
[ "False", "GenerationConfig", "GenerationMixin", "Linear", "ModelCausalLMOutputWithPast", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "True", "__init__", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "cache_position", "can_r...
bart/modeling_bart.py:shift_tokens_right
[ -0.0001357423752779141, 0.026739485561847687, 0.00459761219099164, -0.04174518957734108, -0.0005782272783108056, 0.05189942196011543, 0.04039129242300987, -0.057089366018772125, 0.006572046782821417, 0.007784913759678602, 0.020759768784046173, -0.009872172959148884, -0.0006698974757455289, ...
[ "Model_tokens_right", "Modeled_input_ids", "None", "clone", "decoder_start_token_id", "def", "if", "input_ids", "is", "masked_fill_", "new_zeros", "pad_token_id", "return", "shape" ]
bart/modeling_bart.py:BartLearnedPositionalEmbedding
[ -0.00027780202799476683, 0.014051404781639576, 0.0023514595814049244, -0.002394474111497402, -0.0016847348306328058, 0.023170480504631996, 0.024202829226851463, -0.014452873729169369, 0.002322783228009939, 0.000820860150270164, 0.010782302357256413, 0.02110578492283821, 0.0011112081119790673...
[ "Embedding", "ModelLearnedPositionalEmbedding", "None", "__init__", "arange", "bsz", "class", "def", "device", "dtype", "else", "embedding_dim", "expand", "forward", "if", "input_ids", "is", "long", "nn", "num_embeddings", "offset", "past_key_values_length", "position_ids...
bart/modeling_bart.py:BartScaledWordEmbedding
[ -0.00018138372979592532, 0.03397626802325249, 0.021971320733428, 0.020159251987934113, -0.0008741811034269631, 0.019253218546509743, 0.04462216794490814, -0.020725524052977562, 0.00569102494046092, -0.00651211803779006, 0.005974160507321358, 0.018913455307483673, 0.002505749696865678, -0.0...
[ "Embedding", "ModelScaledWordEmbedding", "__init__", "class", "def", "embed_scale", "embedding_dim", "forward", "input_ids", "nn", "num_embeddings", "padding_idx", "return", "self", "super" ]
bart/modeling_bart.py:eager_attention_forward
[ 0.00002095530362566933, 0.025862814858555794, 0.0269921962171793, -0.01146321278065443, 0.00008955634984886274, 0.03704368323087692, 0.060083046555519104, -0.023942869156599045, 0.021458230912685394, 0.014569009654223919, 0.023152301087975502, 0.026879258453845978, 0.003063444746658206, -0...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "shape", "size", "softmax", "tor...
bart/modeling_bart.py:BartAttention
[ -0.0001932329178089276, 0.04201360046863556, 0.04201360046863556, -0.006776387337595224, -0.0007764610345475376, 0.02010328322649002, 0.029138466343283653, -0.028234947472810745, 0.0020611512009054422, 0.014908052049577236, 0.008188134990632534, 0.009261062368750572, -0.0016235094517469406, ...
[ "ALL_ATTENTION_FUNCTIONS", "EncoderDecoderCache", "False", "Instantiating", "Linear", "ModelAttention", "Module", "None", "Please", "True", "ValueError", "__class__", "__init__", "__name__", "_attn_implementation", "a", "and", "attention_interface", "attention_mask", "attn_outp...
bart/modeling_bart.py:BartEncoderLayer
[ -0.00012346176663413644, 0.036347147077322006, 0.032960765063762665, 0.01275536511093378, -0.0005785065586678684, 0.036798663437366486, 0.02641376480460167, -0.011005735024809837, 0.0042894147336483, 0.010667096823453903, 0.0035698090214282274, 0.022462988272309303, 0.0015027060871943831, ...
[ "ACT2FN", "False", "GradientCheckpointingLayer", "LayerNorm", "Linear", "ModelAttention", "ModelEncoderLayer", "None", "__init__", "activation_dropout", "activation_fn", "activation_function", "and", "any", "attention_dropout", "attention_mask", "attn_weights", "clamp", "clamp_va...
bart/modeling_bart.py:BartDecoderLayer
[ -0.00013334836694411933, 0.051318809390068054, 0.022494366392493248, -0.0071213324554264545, -0.0007629998726770282, 0.0316503643989563, 0.041823696345090866, -0.03504147753119469, 0.007347406353801489, -0.005990962032228708, -0.00477581424638629, 0.02170310728251934, -0.0005157314008101821,...
[ "ACT2FN", "False", "GradientCheckpointingLayer", "LayerNorm", "Linear", "ModelAttention", "ModelDecoderLayer", "None", "True", "__init__", "activation_dropout", "activation_fn", "activation_function", "attention_dropout", "attention_mask", "cache_position", "class", "config", "cr...
bart/modeling_bart.py:BartClassificationHead
[ -0.0002954535011667758, 0.02414887398481369, 0.032806020230054855, 0.012131392024457455, -0.0010394267737865448, 0.019706392660737038, 0.0478421114385128, 0.004157707095146179, -0.003844455350190401, -0.0045848689042031765, 0.01788383722305298, -0.013384399935603142, -0.0007190555334091187, ...
[ "Dropout", "Linear", "ModelClassificationHead", "Module", "__init__", "class", "def", "dense", "dropout", "forward", "hidden_states", "inner_dim", "input_dim", "nn", "num_classes", "out_proj", "p", "pooler_dropout", "return", "self", "super", "tanh", "torch" ]
bart/modeling_bart.py:BartPreTrainedModel
[ -0.0002981277648359537, 0.03427576646208763, 0.007312163710594177, 0.018737420439720154, -0.001271059736609459, 0.02513556368649006, 0.036103807389736176, -0.026963604614138603, -0.001613817410543561, -0.0011425255797803402, 0.005855443887412548, -0.016795126721262932, -0.0026849352288991213...
[ "ModelConfig", "ModelDecoderLayer", "ModelEncoderLayer", "ModelForConditionalGeneration", "ModelPreTrainedModel", "PreTrainedModel", "True", "_can_compile_fullgraph", "_init_weights", "_keys_to_ignore_on_load_unexpected", "_no_split_modules", "_skip_keys_device_placement", "_supports_flash_a...
bart/modeling_bart.py:PretrainedBartModel
[ -0.00029897087370045483, 0.03362969309091568, -0.0044356402941048145, 0.01901817135512829, -0.001369830104522407, 0.024352535605430603, 0.03989177197217941, 0.00420371163636446, 0.006870893761515617, -0.003942791372537613, 0, 0.016119059175252914, -0.002507731318473816, 0.0301507581025362,...
[ "FutureWarning", "ModelModelModel", "The", "__init_subclass__", "been", "class", "def", "depreciated", "has", "instead", "please", "self", "use", "warn", "warnings" ]
bart/modeling_bart.py:BartPretrainedModel
[ -0.0002997910196427256, 0.04114722087979317, 0.0120205357670784, 0.01699056476354599, -0.001545910257846117, 0.02092035673558712, 0.032594144344329834, -0.002109372988343239, -0.005172297824174166, -0.0063859098590910435, 0.005432357545942068, 0.025196893140673637, -0.0045077009126544, 0.0...
[ "FutureWarning", "ModelPreTrainedModel", "ModelPretrainedModel", "PretrainedModelModel", "The", "__init_subclass__", "been", "class", "def", "depreciated", "has", "instead", "please", "self", "use", "warn", "warnings" ]
bart/modeling_bart.py:BartEncoder
[ -0.0002018490486079827, 0.048047155141830444, 0.015524670481681824, 0.010368666611611843, -0.0010057040490210056, 0.04419431835412979, 0.014958076179027557, -0.013938207179307938, 0.003371233120560646, -0.013258295133709908, 0.01745108887553215, 0.016657857224345207, 0, 0.00064450042555108...
[ "BaseModelOutput", "False", "LayerNorm", "ModelEncoder", "ModelEncoderLayer", "ModelLearnedPositionalEmbedding", "ModelPreTrainedModel", "ModelScaledWordEmbedding", "ModuleList", "None", "True", "__init__", "all_attentions", "and", "attention_mask", "attentions", "class", "config",...
bart/modeling_bart.py:BartDecoder
[ -0.0002373179595451802, 0.07217320799827576, 0.00459647411480546, -0.01838589645922184, -0.0014060642570257187, 0.025922972708940506, 0.04979037865996361, -0.03334584832191467, 0.005253113340586424, -0.016101934015750885, 0.000517460226546973, 0.01827169768512249, -0.00229823705740273, 0.0...
[ "BaseModelOutputWithPastAndCrossAttentions", "DynamicCache", "EncoderDecoderCache", "False", "LayerNorm", "ModelDecoder", "ModelDecoderLayer", "ModelLearnedPositionalEmbedding", "ModelPreTrainedModel", "ModelScaledWordEmbedding", "ModuleList", "None", "Setting", "True", "ValueError", "...
bart/modeling_bart.py:BartModel
[ -0.000062998675275594, 0.05443085730075836, 0.004815898835659027, -0.021503549069166183, -0.00013387219223659486, 0.04278310388326645, 0.027439424768090248, -0.040543150156736374, 0.009463801048696041, -0.005935875233262777, 0.027999412268400192, -0.003121934598311782, -0.0016519653145223856...
[ "BaseModelOutput", "If", "ModelDecoder", "ModelEncoder", "ModelModel", "ModelPreTrainedModel", "ModelScaledWordEmbedding", "None", "Please", "Seq2SeqModelOutput", "ValueError", "__init__", "_tied_weights_keys", "and", "are", "attention_mask", "attentions", "auto_docstring", "be",...
bart/modeling_bart.py:BartForConditionalGeneration
[ -0.0003491345851216465, 0.04033489152789116, 0.006961206439882517, -0.00501321442425251, -0.0013464061776176095, 0.03964736685156822, 0.013234886340796947, -0.01982368342578411, -0.004927273839712143, -0.0059299166314303875, 0.026813535019755363, -0.0002309659612365067, 0.0002954215742647648...
[ "CrossEntropyLoss", "False", "GenerationMixin", "Linear", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "Seq2SeqLMOutput", "True", "__init__", "_keys_to_ignore_on_load_missing", "_resize_final_logits_bias", "_tied_weights_keys", "and", "attention_mask",...
bart/modeling_bart.py:BartForSequenceClassification
[ -0.00025072647258639336, 0.034579064697027206, 0.0134474141523242, 0.031640972942113876, -0.0007945557008497417, 0.019888611510396004, 0.016272500157356262, 0.007853741757571697, 0.0030793447513133287, 0.0017021149396896362, 0.04565340653061867, 0.001730365795083344, 0.0017515539657324553, ...
[ "All", "BCEWithLogitsLoss", "CrossEntropyLoss", "False", "MSELoss", "ModelClassificationHead", "ModelForSequenceClassification", "ModelModel", "ModelPreTrainedModel", "None", "NotImplementedError", "Passing", "Seq2SeqSequenceClassifierOutput", "__class__", "__init__", "__name__", "an...
bart/modeling_bart.py:BartForQuestionAnswering
[ -0.00021988233493175358, 0.03354876488447189, 0.011314265429973602, 0.018125340342521667, -0.0006578879547305405, 0.04615769907832146, 0.033999085426330566, 0.012214903719723225, 0.006417045835405588, 0.029270736500620842, 0.02589334361255169, 0.015085686929523945, 0.0024345372803509235, -...
[ "CrossEntropyLoss", "False", "Linear", "ModelForQuestionAnswering", "ModelModel", "ModelPreTrainedModel", "None", "Seq2SeqQuestionAnsweringModelOutput", "__init__", "and", "attention_mask", "auto_docstring", "cache_position", "clamp", "class", "config", "contiguous", "cross_attenti...
bart/modeling_bart.py:BartDecoderWrapper
[ -0.00007960084622027352, 0.058670200407505035, -0.005794242024421692, -0.009013265371322632, -0.0005808238056488335, 0.02396073006093502, 0.03000689670443535, -0.02955903299152851, -0.0018194479634985328, -0.014107719995081425, -0.008117537014186382, 0.00895728264003992, -0.00405876850709319...
[ "ModelDecoder", "ModelDecoderWrapper", "ModelPreTrainedModel", "__init__", "args", "class", "config", "decoder", "def", "forward", "kwargs", "post_init", "return", "self", "super" ]
bart/modeling_bart.py:BartForCausalLM
[ -0.00017654977273195982, 0.051267948001623154, 0.01765146479010582, -0.010287318378686905, -0.0008045751019380987, 0.01787632331252098, 0.027882568538188934, -0.02383509837090969, 0.002881011227145791, 0.010343533009290695, 0.019337909296154976, -0.0021923792082816362, -0.0000619241764070466...
[ "CausalLMOutputWithCrossAttentions", "CrossEntropyLoss", "False", "GenerationMixin", "Linear", "ModelDecoderWrapper", "ModelForCausalLM", "ModelPreTrainedModel", "None", "True", "__init__", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "cache_position", "cla...
swin2sr/modeling_swin2sr.py:Swin2SREncoderOutput
[ -0.00010046965326182544, 0.02211037464439869, 0.028653239831328392, 0.01940298080444336, -0.00066627241903916, 0.03993404284119606, 0.04422074928879738, -0.039482809603214264, 0.015229083597660065, -0.0042867050506174564, 0.0010152722243219614, 0.039482809603214264, -0.0021856555249542, 0....
[ "ModelEncoderOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "last_hidden_state" ]
swin2sr/modeling_swin2sr.py:window_partition
[ -0.00004125074337935075, 0.009773793630301952, -0.008818882517516613, 0.00499923899769783, 0, 0.025501737371087074, 0.019772272557020187, -0.03999391198158264, 0.012245327234268188, 0.007920143194496632, -0.00150258035864681, -0.04403823986649513, 0.002359191421419382, -0.00117959571070969...
[ "Model_partition", "Model_size", "Models", "batch_size", "contiguous", "def", "height", "input_feature", "num_channels", "permute", "return", "shape", "view", "width" ]
swin2sr/modeling_swin2sr.py:window_reverse
[ -0.0000877177226357162, 0.022455736994743347, 0.02098877914249897, 0.02098877914249897, -0.0002803440729621798, 0.03836658596992493, 0.01602369174361229, -0.043783046305179596, 0.017152121290564537, 0.027646509930491447, -0.017942020669579506, -0.0253896526992321, -0.0025530706625431776, 0...
[ "Model_reverse", "Model_size", "Models", "contiguous", "def", "height", "num_channels", "permute", "return", "shape", "view", "width" ]
swin2sr/modeling_swin2sr.py:drop_path
[ 0, 0.017753854393959045, 0.03775503486394882, -0.02123720571398735, 0, 0.043598074465990067, 0.0339345820248127, -0.027304979041218758, 0.014832334592938423, 0.0017486985307186842, 0.020113544538617134, -0.04944111406803131, 0.0016574009787291288, -0.018315685912966728, -0.05034004524350...
[ "False", "Model_path", "Model_prob", "def", "device", "div", "dtype", "floor_", "if", "input", "keep_prob", "ndim", "not", "or", "output", "rand", "random_tensor", "return", "shape", "torch", "training" ]
swin2sr/modeling_swin2sr.py:Swin2SRDropPath
[ -0.00011107319733127952, 0.018094833940267563, 0.03596488758921623, -0.025063030421733856, -0.00047414645086973906, 0.031918834894895554, 0.033717080950737, -0.026861274614930153, 0.0066872211173176765, -0.002430439693853259, 0.009721758775413036, -0.033717080950737, -0.000818341679405421, ...
[ "ModelDropPath", "Module", "None", "__init__", "class", "def", "drop_path", "drop_prob", "extra_repr", "f", "forward", "hidden_states", "nn", "p", "return", "self", "super", "training" ]
swin2sr/modeling_swin2sr.py:Swin2SREmbeddings
[ -0.00007079269562382251, 0.015533939935266972, 0.021162178367376328, 0.02341347374022007, -0.00040277084917761385, 0.019811401143670082, 0.026452723890542984, -0.0030814609490334988, 0.00878005288541317, 0.006022215820848942, 0.01609676331281662, 0.017897799611091614, -0.0008794123423285782,...
[ "Dropout", "ModelEmbeddings", "ModelPatchEmbeddings", "Module", "None", "Parameter", "__init__", "class", "config", "def", "dropout", "else", "embed_dim", "embeddings", "forward", "hidden_dropout_prob", "if", "is", "nn", "not", "num_patches", "output_dimensions", "patch_e...
swin2sr/modeling_swin2sr.py:Swin2SRPatchEmbeddings
[ -0.00010172773181693628, 0.022998394444584846, 0.018714379519224167, 0.03134094923734665, -0.0002906506706494838, 0.005157728213816881, 0.0009089438826777041, -0.0044531202875077724, 0.010484562255442142, 0.014655839651823044, 0.014430364593863487, 0.01815069280564785, -0.0012260173680260777...
[ "Conv2d", "Iterable", "LayerNorm", "ModelPatchEmbeddings", "Module", "None", "True", "_", "__init__", "abc", "class", "collections", "config", "def", "else", "embed_dim", "embeddings", "flatten", "forward", "height", "if", "image_size", "is", "isinstance", "kernel_siz...
swin2sr/modeling_swin2sr.py:Swin2SRPatchUnEmbeddings
[ -0.0002341549698030576, 0.03409010171890259, 0.029285456985235214, 0.024023227393627167, -0.0013370069209486246, 0.01046726293861866, 0.012583594769239426, -0.0329461395740509, 0.007378562353551388, 0.0011082143755629659, -0.03020062856376171, 0.0040038712322711945, -0.004604451823979616, ...
[ "ModelPatchUnEmbeddings", "Module", "__init__", "batch_size", "class", "config", "def", "embed_dim", "embeddings", "forward", "height_width", "nn", "num_channels", "r", "return", "self", "shape", "super", "transpose", "view", "x_size" ]
swin2sr/modeling_swin2sr.py:Swin2SRPatchMerging
[ -0.00019905575027223676, 0.028172966092824936, 0.018744563683867455, 0.01560176257044077, -0.00019993264868389815, 0.013805875554680824, -0.002006341703236103, -0.04018295556306839, 0.0004700171120930463, 0.010270224884152412, 0.004882565699517727, -0.0025394954718649387, -0.0007751774974167...
[ "LayerNorm", "Linear", "ModelPatchMerging", "Module", "__init__", "batch_size", "cat", "class", "def", "dim", "forward", "functional", "height", "if", "input_dimensions", "input_feature", "input_feature_0", "input_feature_1", "input_feature_2", "input_feature_3", "input_resol...
swin2sr/modeling_swin2sr.py:Swin2SRSelfAttention
[ -0.00006184142694110051, 0.04872928932309151, 0.037052225321531296, -0.009712174534797668, -0.0004228023753967136, 0.016841920092701912, 0.023915527388453484, -0.005642043426632881, 0.002484183292835951, 0.024926042184233665, 0.027171632274985313, 0.02335412986576557, 0.0011859518708661199, ...
[ "Dropout", "False", "Iterable", "Linear", "ModelSelfAttention", "Module", "None", "Parameter", "ReLU", "Sequential", "The", "True", "ValueError", "__init__", "a", "abc", "abs", "all_head_size", "arange", "attention", "attention_head_size", "attention_mask", "attention_pro...
swin2sr/modeling_swin2sr.py:Swin2SRSelfOutput
[ -0.00004677825199905783, 0.04611024260520935, 0.04879627376794815, 0.0014549348270520568, -0.00040045680361799896, 0.04051433876156807, 0.03379925340414047, -0.024510055780410767, 0.0021124533377587795, 0.014885102398693562, 0.010744133964180946, 0.011583519168198109, 0.002350279362872243, ...
[ "Dropout", "Linear", "ModelSelfOutput", "Module", "__init__", "attention_probs_dropout_prob", "class", "config", "def", "dense", "dim", "dropout", "forward", "hidden_states", "input_tensor", "nn", "return", "self", "super" ]
swin2sr/modeling_swin2sr.py:Swin2SRAttention
[ 0.00014067025040276349, 0.036236658692359924, 0.04794042184948921, 0.008215142413973808, 0.0005556474789045751, 0.034661151468753815, 0.04073810577392578, -0.01913115382194519, 0.010690939612686634, 0.0015333057381212711, 0.014742243103682995, 0.03038477525115013, 0.0006400496349669993, -0...
[ "False", "Iterable", "ModelAttention", "ModelSelfAttention", "ModelSelfOutput", "Module", "None", "__init__", "abc", "attention_mask", "attention_output", "class", "collections", "config", "def", "dim", "else", "forward", "hidden_states", "if", "isinstance", "nn", "num_he...
swin2sr/modeling_swin2sr.py:Swin2SRIntermediate
[ -0.0002936642267741263, 0.023951541632413864, 0.039881035685539246, 0.0161586944013834, -0.0009311304893344641, 0.04263145104050636, 0.028764769434928894, -0.018565310165286064, -0.004383475985378027, -0.002549865050241351, 0.02097192406654358, -0.025326749309897423, -0.0019768616184592247, ...
[ "ACT2FN", "Linear", "ModelIntermediate", "Module", "__init__", "class", "config", "def", "dense", "dim", "else", "forward", "hidden_act", "hidden_states", "if", "int", "intermediate_act_fn", "isinstance", "mlp_ratio", "nn", "return", "self", "str", "super" ]
swin2sr/modeling_swin2sr.py:Swin2SROutput
[ -0.00027769507141783834, 0.044259220361709595, 0.04999227821826935, 0.024652156978845596, -0.0011824435787275434, 0.03600361570715904, 0.04609379917383194, -0.021097660064697266, -0.0009244558750651777, 0.00905823428183794, 0.007968952879309654, -0.007309651002287865, -0.00014690964599139988...
[ "Dropout", "Linear", "ModelOutput", "Module", "__init__", "class", "config", "def", "dense", "dim", "dropout", "forward", "hidden_dropout_prob", "hidden_states", "int", "mlp_ratio", "nn", "return", "self", "super" ]
swin2sr/modeling_swin2sr.py:Swin2SRLayer
[ -0.00013428596139419824, 0.021008292213082314, 0.02729954570531845, -0.0051397294737398624, -0.00031245622085407376, 0.04740908741950989, 0.01651453971862793, -0.027973609045147896, -0.0007337455172091722, 0.012133131735026836, 0.019884854555130005, -0.0092121921479702, 0.0002931471390184015...
[ "False", "Identity", "Iterable", "LayerNorm", "ModelAttention", "ModelDropPath", "ModelIntermediate", "ModelLayer", "ModelOutput", "Module", "None", "_", "__init__", "_compute_window_shift", "abc", "attention", "attention_output", "attention_outputs", "attention_windows", "attn...
swin2sr/modeling_swin2sr.py:Swin2SRStage
[ -0.00006751780165359378, 0.0020193560048937798, -0.00601570401340723, 0.006439344957470894, 0.0001932862651301548, 0.0424770824611187, 0.01316111721098423, -0.019882889464497566, 0.0020334774162620306, 0.01186195109039545, 0.031179986894130707, -0.026887089014053345, -0.0011155882384628057, ...
[ "Conv2d", "False", "GradientCheckpointingLayer", "LeakyReLU", "ModelLayer", "ModelPatchEmbeddings", "ModelPatchUnEmbeddings", "ModelStage", "ModuleList", "Sequential", "True", "_", "__init__", "class", "config", "conv", "def", "depth", "dim", "drop_path", "elif", "else", ...
swin2sr/modeling_swin2sr.py:Swin2SREncoder
[ -0.00022529016132466495, 0.015843329951167107, 0.006810352206230164, 0.024277908727526665, -0.0007693703519180417, 0.03556200489401817, 0.012309925630688667, -0.03373831510543823, 0.001588607206940651, 0.022454215213656425, 0.023024119436740875, -0.014475560747087002, -0.002735538873821497, ...
[ "False", "ModelEncoder", "ModelEncoderOutput", "ModelStage", "Module", "ModuleList", "None", "True", "__init__", "all_hidden_states", "all_input_dimensions", "all_self_attentions", "attentions", "class", "config", "cpu", "def", "depth", "depths", "device", "dim", "dpr", "...
swin2sr/modeling_swin2sr.py:Swin2SRPreTrainedModel
[ -0.00019061274360865355, 0.04857094958424568, -0.0010166013380512595, 0.023720696568489075, -0.0007165627321228385, 0.020106114447116852, 0.01056135818362236, 0.006240802351385355, 0.0009601234341971576, 0.02744823507964611, 0.006184324622154236, -0.0004730020009446889, -0.001221333513967692...
[ "Conv2d", "LayerNorm", "Linear", "Model", "ModelConfig", "ModelModel", "ModelPreTrainedModel", "ModelSelfAttention", "None", "PreTrainedModel", "True", "_init_weights", "and", "base_model_prefix", "bias", "class", "config", "constant_", "copy_", "create_coords_table_and_index",...
swin2sr/modeling_swin2sr.py:Swin2SRModel
[ -0.000058832814829656854, 0.05125303938984871, 0.010171930305659771, 0.029897606000304222, -0.00010054264566861093, 0.02776206284761429, 0.015173597261309624, -0.016971949487924576, 0.008036387152969837, 0.019332287833094597, 0.021243035793304443, 0.01708434708416462, 0.0014260371681302786, ...
[ "BaseModelOutput", "Conv2d", "False", "LayerNorm", "ModelEmbeddings", "ModelEncoder", "ModelModel", "ModelPatchUnEmbeddings", "ModelPreTrainedModel", "None", "_", "__init__", "and", "attentions", "auto_docstring", "class", "config", "conv_after_body", "def", "else", "embed_di...
swin2sr/modeling_swin2sr.py:Upsample
[ -0.00023062055697664618, 0.02168542891740799, 0.002582950284704566, 0.009707351215183735, -0.0005925174336880445, 0.018619948998093605, 0.019641775637865067, -0.06403446197509766, 0.004882059991359711, 0.009707351215183735, 0.029065286740660667, -0.047458160668611526, 0.0012488990323618054, ...
[ "Conv2d", "Model", "Module", "PixelShuffle", "Scale", "Supported", "ValueError", "__getattr__", "__init__", "add_module", "class", "convolution", "convolution_", "def", "elif", "else", "f", "for", "forward", "hidden_state", "i", "if", "in", "int", "is", "log2", "m...
swin2sr/modeling_swin2sr.py:UpsampleOneStep
[ -0.0002281921188114211, 0.00707572465762496, -0.0013160847593098879, 0.013019333593547344, -0.0007960190414451063, 0.026717936620116234, 0.025359397754073143, -0.047548867762088776, 0.005207733251154423, 0.023095164448022842, 0.016415681689977646, -0.044831790030002594, 0.0003714755293913185...
[ "Conv2d", "ModelOneStep", "Module", "PixelShuffle", "__init__", "class", "conv", "def", "forward", "in_channels", "nn", "out_channels", "pixel_shuffle", "return", "scale", "self", "super", "x" ]
swin2sr/modeling_swin2sr.py:PixelShuffleUpsampler
[ -0.00046592229045927525, 0.032294124364852905, 0.0271082054823637, 0.03936583176255226, -0.001900520990602672, 0.041723065078258514, -0.00013075290189590305, -0.023808076977729797, -0.008014599792659283, 0.009723595343530178, 0.017443541437387466, -0.01013611163944006, 0.0002578226267360151,...
[ "Conv2d", "LeakyReLU", "ModelShuffleUpsampler", "Module", "True", "Upsample", "__init__", "activation", "class", "config", "conv_before_upsample", "def", "embed_dim", "final_convolution", "forward", "inplace", "nn", "num_channels_out", "num_features", "return", "self", "seq...
swin2sr/modeling_swin2sr.py:NearestConvUpsampler
[ -0.00038542484981007874, 0.035604529082775116, 0.034906402230262756, 0.03397556394338608, -0.0016435097204521298, 0.027692411094903946, 0.010122856125235558, -0.025830736383795738, -0.005672290455549955, -0.0052068717777729034, 0.01687142811715603, 0.003810615511611104, 0.0007344888872466981...
[ "Conv2d", "LeakyReLU", "Model", "ModelConvUpsampler", "Module", "The", "True", "ValueError", "__init__", "activation", "an", "at", "class", "config", "conv", "conv_before_upsample", "conv_hr", "conv_up1", "conv_up2", "def", "embed_dim", "factor", "final_convolution", "f...
swin2sr/modeling_swin2sr.py:PixelShuffleAuxUpsampler
[ -0.0005523759755305946, 0.02212509885430336, 0.029700541868805885, 0.01972019672393799, -0.0022996875923126936, 0.045452650636434555, 0.0005260723410174251, -0.0435287281870842, -0.0062226844020187855, 0.00907850544899702, 0.0070944614708423615, 0.01226500142365694, -0.0017435540212318301, ...
[ "Conv2d", "LeakyReLU", "ModelShuffleAuxUpsampler", "Module", "Sequential", "True", "Upsample", "__init__", "activation", "aux", "bicubic", "class", "config", "conv_after_aux", "conv_aux", "conv_before_upsample", "conv_bicubic", "def", "embed_dim", "final_convolution", "forwar...
swin2sr/modeling_swin2sr.py:Swin2SRForImageSuperResolution
[ -0.0003642925585154444, 0.06232649087905884, 0.0201983992010355, 0.0373958945274353, -0.0015148799866437912, 0.02366098202764988, 0.01050316821783781, -0.0302398893982172, -0.00819477904587984, 0.025738531723618507, 0.01638955809175968, 0.003953115548938513, 0.00104598852340132, 0.02296846...
[ "Conv2d", "False", "ImageSuperResolutionOutput", "Model", "ModelForImageSuperResolution", "ModelModel", "ModelPreTrainedModel", "NearestConvUpsampler", "None", "PixelShuffleAuxUpsampler", "PixelShuffleUpsampler", "UpsampleOneStep", "__init__", "align_corners", "attentions", "auto_docst...
cohere/modeling_cohere.py:CohereLayerNorm
[ -0.000010297008884663228, 0.036567963659763336, 0.038811396807432175, 0.047560784965753555, -0.00008807228732621297, 0.030735036358237267, 0.018283981829881668, -0.014918831177055836, 0.009366334415972233, 0.04733644425868988, 0.016601406037807465, 0.006169441621750593, 0.003477321704849601,...
[ "ModelLayerNorm", "Module", "None", "Parameter", "True", "__init__", "class", "def", "eps", "float32", "forward", "hidden_size", "hidden_states", "keepdim", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "super", "to", "torch", "variance", "variance_epsilon...
cohere/modeling_cohere.py:CohereRotaryEmbedding
[ -0.00029855233151465654, 0.048173680901527405, 0.002417369047179818, -0.005037392023950815, -0.0014764770166948438, 0.04215196892619133, 0.041225552558898926, 0.004689985420554876, -0.002402893966063857, 0.0207285787910223, 0.0015271403826773167, -0.0037925192154943943, -0.001179734012112021...
[ "False", "ModelRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "arange", "attention_factor", "attention_scaling", "base", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cpu", "def", "default", "device", "device...
cohere/modeling_cohere.py:CohereMLP
[ -0.0002434849739074707, 0.024471141397953033, 0.0228551235049963, 0.026548879221081734, -0.0009450823999941349, 0.060485273599624634, 0.036244992166757584, -0.005078915972262621, 0, -0.005021201446652412, 0.028164898976683617, -0.04640282690525055, -0.0015583038330078125, 0.009176678024232...
[ "ACT2FN", "Linear", "ModelMLP", "Module", "__init__", "act_fn", "class", "config", "def", "down_proj", "forward", "gate_proj", "hidden_act", "hidden_size", "intermediate_size", "nn", "return", "self", "super", "up_proj", "x" ]
cohere/modeling_cohere.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
cohere/modeling_cohere.py:eager_attention_forward
[ 0, 0.021594731137156487, 0.01775064319372177, -0.018768195062875748, -0.00008832923776935786, 0.039797618985176086, 0.05359111353754997, -0.035049039870500565, 0.02069023996591568, 0.011645326390862465, 0.029395969584584236, 0.022499222308397293, 0.002741739386692643, -0.014584923163056374...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "causal_mask", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "key_states", "kwargs", "matmul", "module", "nn", "not", "num_key_value_groups", ...
cohere/modeling_cohere.py:rotate_half
[ 0.00004134093614993617, 0.011483984999358654, 0.030173607170581818, -0.007937460206449032, 0.0003535969590302557, 0.03129948675632477, 0.01598750799894333, -0.02792184427380562, 0.018577033653855324, 0.03422677889466286, -0.00529163982719183, 0.0008197819697670639, -0.0004046134417876601, ...
[ "Model_half", "def", "dim", "flatten", "return", "rot_x", "stack", "torch", "x", "x1", "x2" ]
cohere/modeling_cohere.py:apply_rotary_pos_emb
[ -0.00014119588013272732, 0.029553454369306564, 0.02216508984565735, -0.000987483188509941, -0.0007423884235322475, 0.02387009747326374, 0.04978620260953903, -0.006393775809556246, 0.012673884630203247, 0.03773748502135277, 0.006308525800704956, 0.002358592813834548, -0.0006713464972563088, ...
[ "Model_rotary_pos_emb", "cos", "def", "float", "k", "k_embed", "q", "q_embed", "return", "rotate_half", "sin", "unsqueeze", "unsqueeze_dim" ]
cohere/modeling_cohere.py:CohereAttention
[ -0.00008214123226935044, 0.04003221169114113, 0.018441805616021156, 0.0019397629657760262, -0.0002003016124945134, 0.02035345509648323, 0.034634608775377274, -0.01383135374635458, 0.00047088449355214834, 0.025188807398080826, 0.019453855231404305, 0.0317109078168869, -0.00008258048910647631,...
[ "ALL_ATTENTION_FUNCTIONS", "Linear", "ModelAttention", "ModelLayerNorm", "Module", "None", "Tensor", "True", "__init__", "_attn_implementation", "apply_rotary_pos_emb", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "cache_kwargs", "cac...
cohere/modeling_cohere.py:CohereDecoderLayer
[ -0.00014439009828493, 0.040119316428899765, 0.01983426883816719, -0.006536292843520641, -0.0007606892613694072, 0.03380841016769409, 0.043500155210494995, -0.036513086408376694, 0.0040006618946790695, -0.008508450351655483, -0.0023243282921612263, 0.015889953821897507, -0.0019721572753041983...
[ "False", "GradientCheckpointingLayer", "ModelAttention", "ModelDecoderLayer", "ModelLayerNorm", "ModelMLP", "None", "Tensor", "_", "__init__", "attention_mask", "cache_position", "class", "config", "def", "eps", "forward", "hidden_size", "hidden_states", "hidden_states_attentio...
cohere/modeling_cohere.py:CoherePreTrainedModel
[ -0.00034975787275470793, 0.027630871161818504, 0.016904963180422783, 0.011250545270740986, -0.0018799485405907035, 0.029612833634018898, 0.03077869303524494, -0.03241089731454849, -0.00408050836995244, 0.003993069287389517, 0.006383080966770649, 0.004459412768483162, -0.004255387466400862, ...
[ "ModelAttention", "ModelConfig", "ModelDecoderLayer", "ModelPreTrainedModel", "PreTrainedModel", "True", "_can_compile_fullgraph", "_can_record_outputs", "_no_split_modules", "_skip_keys_device_placement", "_supports_attention_backend", "_supports_flash_attn", "_supports_flex_attn", "_supp...
cohere/modeling_cohere.py:CohereModel
[ -0.00011140928836539388, 0.04644490405917168, -0.008285680785775185, -0.009356618858873844, -0.000852523255161941, 0.038102857768535614, 0.04035746678709984, -0.01533132791519165, 0.010709383524954319, 0.0075529334135353565, 0.015894979238510132, -0.0024237025063484907, -0.002000963781028986...
[ "BaseModelOutputWithPast", "DynamicCache", "Embedding", "False", "ModelDecoderLayer", "ModelLayerNorm", "ModelModel", "ModelPreTrainedModel", "ModelRotaryEmbedding", "ModuleList", "None", "ValueError", "You", "__init__", "and", "arange", "attention_mask", "auto_docstring", "cache...
cohere/modeling_cohere.py:CohereForCausalLM
[ -0.0003332803607918322, 0.0422036275267601, 0.012318897061049938, -0.0025521789211779833, -0.0012689604191109538, 0.025436241179704666, 0.03650043532252312, -0.003735591424629092, 0.0002299099724041298, 0.020987750962376595, 0.029656603932380676, 0.01226186566054821, 0.0013901533093303442, ...
[ "CausalLMOutputWithPast", "GenerationMixin", "Linear", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "_pp_plan", "_tied_weights_keys", "_tp_plan", "attention_mask", "attentions", "auto_docstring", "cache_position", "can_return_tuple", "class", "colwi...
pegasus_x/modeling_pegasus_x.py:DimensionInfo
[ -0.0001596910151420161, -0.007267715409398079, -0.006785093806684017, -0.006898651830852032, -0.0007807116489857435, 0.028048839420080185, 0.059958651661872864, 0.002242771675810218, 0.01283206045627594, 0.012321049347519875, 0.012775281444191933, 0.01447865180671215, -0.0032505993731319904,...
[ "ModelInfo", "batch_size", "block_size", "class", "dim_per_head", "global_len", "hidden_dim", "int", "num_blocks", "num_heads", "padded_seq_len", "seq_len" ]
pegasus_x/modeling_pegasus_x.py:shift_tokens_right
[ -0.0001357423752779141, 0.026739485561847687, 0.00459761219099164, -0.04174518957734108, -0.0005782272783108056, 0.05189942196011543, 0.04039129242300987, -0.057089366018772125, 0.006572046782821417, 0.007784913759678602, 0.020759768784046173, -0.009872172959148884, -0.0006698974757455289, ...
[ "Model_tokens_right", "Modeled_input_ids", "None", "clone", "decoder_start_token_id", "def", "if", "input_ids", "is", "masked_fill_", "new_zeros", "pad_token_id", "return", "shape" ]
pegasus_x/modeling_pegasus_x.py:PegasusXScaledWordEmbedding
[ -0.00018138372979592532, 0.03397626802325249, 0.021971320733428, 0.020159251987934113, -0.0008741811034269631, 0.019253218546509743, 0.04462216794490814, -0.020725524052977562, 0.00569102494046092, -0.00651211803779006, 0.005974160507321358, 0.018913455307483673, 0.002505749696865678, -0.0...
[ "Embedding", "ModelScaledWordEmbedding", "__init__", "class", "def", "embed_scale", "embedding_dim", "forward", "input_ids", "nn", "num_embeddings", "padding_idx", "return", "self", "super" ]
pegasus_x/modeling_pegasus_x.py:PegasusXSinusoidalPositionalEmbedding
[ -0.0002799366775434464, 0.042492952197790146, 0.008211475796997547, 0.015159647911787033, -0.0015288849826902151, 0.003387951757758856, 0.049383703619241714, 0, 0.00849859043955803, 0.009187665767967701, 0.018375331535935402, 0.007981784641742706, 0.000509628327563405, -0.00559873366728425...
[ "ModelSinusoidalPositionalEmbedding", "Module", "None", "__init__", "arange", "batch_size", "class", "cos", "def", "device", "div_term", "dtype", "embed_dim", "exp", "expand", "float", "forward", "half_d_feature", "if", "input_embeds", "int64", "is", "log", "long", "m...
pegasus_x/modeling_pegasus_x.py:eager_attention_forward
[ 0.00002095530362566933, 0.025862814858555794, 0.0269921962171793, -0.01146321278065443, 0.00008955634984886274, 0.03704368323087692, 0.060083046555519104, -0.023942869156599045, 0.021458230912685394, 0.014569009654223919, 0.023152301087975502, 0.026879258453845978, 0.003063444746658206, -0...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "shape", "size", "softmax", "tor...
pegasus_x/modeling_pegasus_x.py:PegasusXAttention
[ -0.0001932329178089276, 0.04201360046863556, 0.04201360046863556, -0.006776387337595224, -0.0007764610345475376, 0.02010328322649002, 0.029138466343283653, -0.028234947472810745, 0.0020611512009054422, 0.014908052049577236, 0.008188134990632534, 0.009261062368750572, -0.0016235094517469406, ...
[ "ALL_ATTENTION_FUNCTIONS", "EncoderDecoderCache", "False", "Instantiating", "Linear", "ModelAttention", "Module", "None", "Please", "True", "ValueError", "__class__", "__init__", "__name__", "_attn_implementation", "a", "and", "attention_interface", "attention_mask", "attn_outp...
pegasus_x/modeling_pegasus_x.py:PegasusXGlobalLocalAttention
[ -0.0001223433209815994, 0.02951730787754059, 0.02839069254696369, -0.027038754895329475, -0.0005210593226365745, 0.028278030455112457, 0.031094567850232124, -0.03357312083244324, -0.0007358202710747719, 0.0066188620403409, 0.03334779664874077, 0.02861601486802101, -0.0011970282066613436, -...
[ "BHGF", "BHGX", "BHNKF", "BHNKG", "BHNKX", "BHNXF", "BHXF", "DimensionInfo", "False", "Linear", "ModelGlobalLocalAttention", "Module", "None", "ValueError", "__init__", "_shape", "and", "attention_mask", "attn_output", "attn_probs", "attn_weights", "batch_size", "be", "...
pegasus_x/modeling_pegasus_x.py:PegasusXEncoderLayer
[ -0.0001452143769711256, 0.013349161483347416, 0.016447069123387337, -0.02151636965572834, -0.00030979065923020244, 0.042356833815574646, 0.018925393000245094, -0.04123032093048096, 0.0019432323751971126, -0.0035203485749661922, 0.015320557169616222, 0.027261579409241676, 0.000869526120368391...
[ "ACT2FN", "False", "GradientCheckpointingLayer", "LayerNorm", "Linear", "ModelEncoderLayer", "ModelGlobalLocalAttention", "__init__", "activation_dropout", "activation_fn", "activation_function", "attention_dropout", "attention_mask", "attn_weights", "block_size", "class", "classmeth...
pegasus_x/modeling_pegasus_x.py:PegasusXDecoderLayer
[ -0.00014653550169896334, 0.05242793262004852, 0.02101636864244938, -0.008191864006221294, -0.000674416427500546, 0.02937772125005722, 0.04158077389001846, -0.029603702947497368, 0.005903792101889849, -0.006722978316247463, -0.003898197552189231, 0.021129360422492027, -0.0001588939194334671, ...
[ "ACT2FN", "False", "GradientCheckpointingLayer", "LayerNorm", "Linear", "ModelAttention", "ModelDecoderLayer", "None", "True", "__init__", "activation_dropout", "activation_fn", "activation_function", "attention_dropout", "attention_mask", "cache_position", "class", "config", "cr...
pegasus_x/modeling_pegasus_x.py:PegasusXPreTrainedModel
[ -0.00027046079048886895, 0.029804421588778496, 0.02178015373647213, 0.012724194675683975, -0.0015332081820815802, 0.02911662682890892, 0.029804421588778496, -0.029575156047940254, -0.0024932543747127056, 0.005273089744150639, 0.006362097337841988, 0.0026795321609824896, -0.004012133460491896...
[ "False", "ModelConfig", "ModelDecoderLayer", "ModelEncoderLayer", "ModelPreTrainedModel", "PreTrainedModel", "True", "_can_compile_fullgraph", "_no_split_modules", "_supports_flash_attn", "_supports_flex_attn", "_supports_sdpa", "base_model_prefix", "class", "config", "model", "r", ...
pegasus_x/modeling_pegasus_x.py:PegasusXEncoder
[ -0.00016553018940612674, 0.03988303989171982, 0.004277229309082031, -0.006968201603740454, -0.0006585800438188016, 0.04781432822346687, 0.01597587764263153, -0.025493420660495758, 0.0021386146545410156, -0.010027412325143814, 0.020168127492070198, 0.025493420660495758, -0.0002832602185662836...
[ "BaseModelOutput", "Embedding", "False", "LayerNorm", "ModelEncoder", "ModelEncoderLayer", "ModelPreTrainedModel", "ModelScaledWordEmbedding", "ModelSinusoidalPositionalEmbedding", "ModuleList", "None", "Setting", "True", "_", "__init__", "all_attentions", "and", "arange", "atten...
pegasus_x/modeling_pegasus_x.py:PegasusXDecoder
[ -0.0002230503159807995, 0.0721755102276802, 0.006880655884742737, -0.019642703235149384, -0.0013846963411197066, 0.024895982816815376, 0.052304405719041824, -0.03517414256930351, 0.006052693352103233, -0.011705679818987846, -0.0023268607910722494, 0.012847697362303734, -0.0024838882964104414...
[ "BaseModelOutputWithPastAndCrossAttentions", "DynamicCache", "EncoderDecoderCache", "False", "LayerNorm", "ModelDecoder", "ModelDecoderLayer", "ModelPreTrainedModel", "ModelScaledWordEmbedding", "ModelSinusoidalPositionalEmbedding", "ModuleList", "None", "Setting", "True", "ValueError", ...
pegasus_x/modeling_pegasus_x.py:PegasusXModel
[ -0.00010636501247063279, 0.059165213257074356, 0.006359139923006296, -0.022299010306596756, -0.00031865734490565956, 0.036305926740169525, 0.02353161759674549, -0.036978255957365036, 0.004706323612481356, -0.00423008855432272, 0.02218695357441902, 0.007843872532248497, -0.002129051135852933,...
[ "BaseModelOutput", "ModelDecoder", "ModelEncoder", "ModelModel", "ModelPreTrainedModel", "ModelScaledWordEmbedding", "None", "Seq2SeqModelOutput", "__init__", "_tied_weights_keys", "and", "attention_mask", "attentions", "auto_docstring", "cache_position", "class", "config", "cross_...
pegasus_x/modeling_pegasus_x.py:PegasusXForConditionalGeneration
[ -0.0003062985197175294, 0.036242950707674026, 0.0001202043640660122, -0.0082629369571805, -0.0011610850924625993, 0.03783855214715004, 0.013277684338390827, -0.017779560759663582, -0.004330918658524752, -0.0001602724805707112, 0.0183494184166193, 0.003148463787510991, -0.0007051989086903632,...
[ "CrossEntropyLoss", "False", "GenerationMixin", "Linear", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "Seq2SeqLMOutput", "__init__", "_tied_weights_keys", "and", "attention_mask", "auto_docstring", "base_model_prefix", "cache_position", "class", "...
pegasus_x/modeling_pegasus_x.py:PegasusXDecoderWrapper
[ -0.00007960084622027352, 0.058670200407505035, -0.005794242024421692, -0.009013265371322632, -0.0005808238056488335, 0.02396073006093502, 0.03000689670443535, -0.02955903299152851, -0.0018194479634985328, -0.014107719995081425, -0.008117537014186382, 0.00895728264003992, -0.00405876850709319...
[ "ModelDecoder", "ModelDecoderWrapper", "ModelPreTrainedModel", "__init__", "args", "class", "config", "decoder", "def", "forward", "kwargs", "post_init", "return", "self", "super" ]
xlm_roberta/modeling_xlm_roberta.py:XLMRobertaEmbeddings
[ -0.00023229305224958807, 0.011745871976017952, 0.008171041496098042, -0.011632384732365608, -0.0012341677211225033, 0.025307530537247658, 0.02190292999148369, -0.014356065541505814, 0.00023229305224958807, -0.01055426150560379, 0.024059176445007324, 0.0160016231238842, -0.0004238018300384283...
[ "Dropout", "Embedding", "False", "LayerNorm", "Model", "Module", "None", "__init__", "arange", "batch_size", "buffered_token_type_ids", "class", "config", "create_position_ids_from_input_ids", "create_position_ids_from_inputs_embeds", "cumsum", "def", "device", "dim", "dropout"...
xlm_roberta/modeling_xlm_roberta.py:eager_attention_forward
[ 0.00002095530362566933, 0.025862814858555794, 0.0269921962171793, -0.01146321278065443, 0.00008955634984886274, 0.03704368323087692, 0.060083046555519104, -0.023942869156599045, 0.021458230912685394, 0.014569009654223919, 0.023152301087975502, 0.026879258453845978, 0.003063444746658206, -0...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "shape", "size", "softmax", "tor...