identifier
stringlengths
24
102
embedding
listlengths
2.56k
2.56k
tokens
listlengths
4
448
internvl/modeling_internvl.py:InternVLVisionPreTrainedModel
[ -0.00014443512191064656, 0.04369819164276123, -0.010364314541220665, 0.024874355643987656, -0.0006162565550766885, 0.023977981880307198, 0.025658681988716125, -0.02756347507238388, -0.002423008671030402, 0.022521374747157097, 0.019832255318760872, 0.008123381994664669, -0.001561650075018406,...
[ "ModelVisionAttention", "ModelVisionConfig", "ModelVisionEmbeddings", "ModelVisionLayer", "ModelVisionPreTrainedModel", "Model_vision", "None", "PreTrainedModel", "True", "_can_record_outputs", "_init_weights", "_no_split_modules", "_supports_attention_backend", "_supports_flash_attn", "...
internvl/modeling_internvl.py:InternVLVisionModel
[ 0, 0.042341239750385284, 0.00168720877263695, 0.051526375114917755, 0.00018114740669261664, 0.026211243122816086, 0.027667423710227013, -0.020162494853138924, 0.010697323828935623, 0.014561801217496395, 0.027779437601566315, 0.022290756925940514, -0.00028528529219329357, -0.013049614615738...
[ "False", "Identity", "LayerNorm", "ModelVisionEmbeddings", "ModelVisionEncoder", "ModelVisionModel", "ModelVisionModelOutputWithPooling", "ModelVisionPreTrainedModel", "None", "__init__", "attentions", "auto_docstring", "bool_masked_pos", "check_model_inputs", "class", "config", "def...
internvl/modeling_internvl.py:InternVLPreTrainedModel
[ -0.0004097087075933814, 0.03268850967288017, 0.012934301979839802, 0.019283868372440338, -0.00199893768876791, 0.032218169420957565, 0.029984064400196075, -0.035040199756622314, -0.005026740022003651, 0.017402514815330505, 0.017520099878311157, 0.008818842470645905, -0.004762175027281046, ...
[ "ModelConfig", "ModelPreTrainedModel", "PreTrainedModel", "True", "_can_compile_fullgraph", "_skip_keys_device_placement", "_supports_attention_backend", "_supports_flash_attn", "_supports_flex_attn", "_supports_sdpa", "base_model_prefix", "class", "config", "image", "input_modalities", ...
internvl/modeling_internvl.py:InternVLMultiModalProjector
[ -0.00024785782443359494, 0.046455640345811844, 0.043736282736063004, 0.04350966960191727, -0.0006019404390826821, 0.031499188393354416, 0.02492741495370865, -0.031499188393354416, -0.0004709298664238304, 0.022208061069250107, 0.027306850999593735, -0.004418950993567705, -0.001784576335921883...
[ "ACT2FN", "LayerNorm", "Linear", "ModelMultiModalProjector", "Module", "__init__", "act", "class", "config", "def", "downsample_ratio", "forward", "hidden_size", "hidden_states", "image_features", "int", "layer_norm", "linear_1", "linear_2", "nn", "projector_hidden_act", "r...
internvl/modeling_internvl.py:InternVLModelOutputWithPast
[ -0.000123380494187586, 0.02538112923502922, 0.014777458272874355, 0.01703355833888054, -0.0007931602885946631, 0.02910369634628296, 0.04331712797284126, -0.038353707641363144, 0.016920752823352814, -0.007952754385769367, 0.02763723023235798, 0.026170765981078148, -0.002354804892092943, 0.0...
[ "BaseModelOutputWithPast", "ModelModelOutputWithPast", "None", "class", "image_hidden_states", "r" ]
internvl/modeling_internvl.py:InternVLModel
[ -0.00016251362103503197, 0.045629628002643585, 0.0023905227426439524, 0.01912418194115162, -0.0005312272696755826, 0.03310384973883629, 0.01890050806105137, -0.04965577274560928, -0.00349491648375988, 0.02952505461871624, 0.027623819187283516, 0.005675744265317917, 0.0008597494452260435, 0...
[ "AutoModel", "False", "Height", "Image", "ModelModel", "ModelModelOutputWithPast", "ModelMultiModalProjector", "ModelPreTrainedModel", "None", "Obtains", "True", "ValueError", "You", "__init__", "_checkpoint_conversion_mapping", "all", "and", "apply", "attention_mask", "attenti...
internvl/modeling_internvl.py:InternVLCausalLMOutputWithPast
[ -0.00020158913685008883, 0.02637176588177681, 0.02978190779685974, 0.009434727020561695, -0.0011793408775702119, 0.03387407958507538, 0.03773890808224678, -0.030463937669992447, 0.01784641109406948, -0.004205842036753893, 0.020006166771054268, 0.018642110750079155, -0.0008596400148235261, ...
[ "ModelCausalLMOutputWithPast", "ModelOutput", "None", "attentions", "class", "hidden_states", "image_hidden_states", "logits", "loss", "past_key_values", "r" ]
internvl/modeling_internvl.py:InternVLForConditionalGeneration
[ -0.0002781958901323378, 0.038557592779397964, 0.005783638916909695, 0.02245412766933441, -0.0009922910248860717, 0.03152650222182274, 0.02619648352265358, -0.03719673678278923, -0.0032603847794234753, 0.025856269523501396, 0.03515545278787613, 0.001148222479969263, 0.0025374300312250853, 0...
[ "False", "GenerationMixin", "Linear", "ModelCausalLMOutputWithPast", "ModelForConditionalGeneration", "ModelModel", "ModelPreTrainedModel", "None", "True", "__init__", "_checkpoint_conversion_mapping", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "cache_posit...
d_fine/modeling_d_fine.py:multi_scale_deformable_attention_v2
[ -0.00027023896109312773, 0.0375685840845108, 0.0066145905293524265, -0.01924244500696659, -0.0011024317936971784, 0.010823875665664673, 0.055894721299409866, -0.06459964066743851, 0.0044669960625469685, 0.026916515082120895, 0.026343824341893196, 0.014603640884160995, -0.0012742392718791962,...
[ "False", "Model_scale_deformable_attention_v2", "_", "align_corners", "append", "arange", "attention_weights", "batch_size", "bilinear", "clamp", "concat", "contiguous", "def", "default", "device", "dim", "discrete", "elif", "enumerate", "flatten", "for", "functional", "g...
d_fine/modeling_d_fine.py:DFineMultiscaleDeformableAttention
[ -0.00031688500894233584, 0.05277575924992561, 0.020050179213285446, -0.00087503477698192, -0.0011523091234266758, 0.03157326951622963, 0.047705598175525665, -0.05484991520643234, 0.0030680231284350157, 0.007201932370662689, 0.00021425748127512634, 0.02684880420565605, -0.0018004830926656723,...
[ "F", "Last", "Linear", "Make", "Model", "Module", "None", "ValueError", "_", "__init__", "align", "attention_mask", "attention_weights", "batch_size", "be", "but", "class", "config", "d_model", "decoder_attention_heads", "decoder_method", "decoder_n_points", "decoder_offs...
d_fine/modeling_d_fine.py:DFineGate
[ -0.00008731099660508335, 0.012869112193584442, 0.016029946506023407, 0.006970769260078669, -0.00018696901679504663, 0.042671266943216324, 0.019303668290376663, -0.044025909155607224, 0.012474007904529572, 0.012022459879517555, 0.023819146677851677, -0.0282217375934124, 0.003118501976132393, ...
[ "LayerNorm", "Linear", "Model", "Module", "__init__", "cat", "chunk", "class", "d_model", "def", "dim", "forward", "gate", "gate1", "gate2", "gate_input", "gates", "hidden_states", "nn", "norm", "return", "second_residual", "self", "sigmoid", "super", "torch" ]
d_fine/modeling_d_fine.py:DFineMultiheadAttention
[ -0.00019584524852689356, 0.04809463024139404, 0.03765900060534477, -0.011229642666876316, -0.0008365516550838947, 0.024387607350945473, 0.025408484041690826, -0.0037715716753154993, -0.00208428967744112, 0.022005561739206314, 0.004310367628931999, 0.002282793400809169, -0.0030342719983309507...
[ "Attention", "False", "Linear", "Model", "Module", "None", "ValueError", "__init__", "_reshape", "and", "attention_mask", "attn_output", "attn_probs", "attn_weights", "attn_weights_reshaped", "batch_size", "be", "bmm", "bool", "but", "by", "class", "contiguous", "def", ...
d_fine/modeling_d_fine.py:DFineDecoderLayer
[ -0.00014995847595855594, 0.054971762001514435, 0.017604593187570572, -0.005252983421087265, -0.0005927353049628437, 0.030893221497535706, 0.039752308279275894, -0.015787344425916672, 0.006729497574269772, -0.011017068289220333, 0.0035777075681835413, 0.009824498556554317, 0.00037800185964442...
[ "ACT2FN", "False", "LayerNorm", "Linear", "Model", "ModelGate", "ModelMultiheadAttention", "ModelMultiscaleDeformableAttention", "Module", "None", "__init__", "activation_dropout", "activation_fn", "attention_dropout", "attention_mask", "clamp", "class", "config", "cross_attn_wei...
d_fine/modeling_d_fine.py:DFinePreTrainedModel
[ -0.0001962610549526289, 0.045962050557136536, -0.02433284930884838, 0.016221899539232254, -0.0006301475805230439, 0.025572022423148155, 0.0157712921500206, -0.0344715379178524, 0.003689355682581663, 0.010420317761600018, 0.004139964003115892, 0.012673359364271164, -0.00041540456004440784, ...
[ "BatchNorm2d", "Conv2d", "LayerNorm", "Linear", "Model", "ModelConfig", "ModelDecoder", "ModelDecoderLayer", "ModelForObjectDetection", "ModelGate", "ModelHybridEncoder", "ModelLQE", "ModelModel", "ModelMultiscaleDeformableAttention", "None", "PreTrainedModel", "True", "_", "_ini...
d_fine/modeling_d_fine.py:DFineIntegral
[ -0.0002754678134806454, 0.034338053315877914, 0.017284253612160683, 0.0259263813495636, -0.0010874676518142223, 0.026963436976075172, 0.07098066806793213, -0.045399975031614304, 0.011349993757903576, 0.014691616408526897, 0.025350239127874374, -0.01981927827000618, -0.0017212236998602748, ...
[ "F", "Model", "Module", "_", "__init__", "batch_size", "class", "config", "def", "device", "dim", "forward", "linear", "max_num_bins", "nn", "num_queries", "pred_corners", "project", "reshape", "return", "self", "shape", "softmax", "super", "to" ]
d_fine/modeling_d_fine.py:DFineDecoderOutput
[ -0.00011808373528765514, 0.026820100843906403, 0.023751700296998024, 0.0260245893150568, -0.0007351378444582224, 0.05909514054656029, 0.06591381132602692, -0.01625116355717182, 0.01545565202832222, -0.010398471727967262, 0.0060799806378781796, 0.016819385811686516, -0.002912140218541026, 0...
[ "Model", "ModelOutput", "None", "attentions", "class", "cross_attentions", "hidden_states", "initial_reference_points", "intermediate_hidden_states", "intermediate_logits", "intermediate_predicted_corners", "intermediate_reference_points", "last_hidden_state", "r" ]
d_fine/modeling_d_fine.py:inverse_sigmoid
[ 0.00006263991963351145, 0.023037374019622803, 0.018407313153147697, 0.016374604776501656, 0.00035466544795781374, 0.03929905220866203, 0.036588769406080246, 0.015019464306533337, 0.015132392756640911, 0.013777253217995167, 0.043590325862169266, -0.041105903685092926, 0.0016304024029523134, ...
[ "Model_sigmoid", "clamp", "def", "eps", "log", "max", "min", "return", "torch", "x", "x1", "x2" ]
d_fine/modeling_d_fine.py:weighting_function
[ -0.00024668656988069415, -0.004833626560866833, -0.00047549730516038835, 0.01738961599767208, -0.0008616154082119465, 0.05949079245328903, 0.03981306776404381, -0.0007329093641601503, 0.006063484586775303, 0.025741208344697952, 0.02516918070614338, -0.07504992187023163, 0.0004415332223288715...
[ "Model_function", "None", "abs", "cat", "def", "for", "i", "in", "left_values", "max_num_bins", "range", "reg_scale", "return", "right_values", "step", "torch", "up", "upper_bound1", "upper_bound2", "values", "zeros_like" ]
d_fine/modeling_d_fine.py:distance2bbox
[ -0.00009204662637785077, -0.03262699022889137, -0.008779832161962986, -0.029454920440912247, -0.00039296830072999, 0.07386387884616852, 0.02435695379972458, -0.030814379453659058, 0.012461697682738304, 0.016426783055067062, 0.002294085221365094, -0.013198070228099823, -0.0001938289642566815,...
[ "Model", "abs", "bboxes", "bottom_right_x", "bottom_right_y", "corners_to_center_format", "def", "distance", "points", "reg_scale", "return", "stack", "top_left_x", "top_left_y", "torch" ]
d_fine/modeling_d_fine.py:DFineDecoder
[ -0.00016439262253697962, 0.03280781954526901, -0.0039878468960523605, 0.016290778294205666, -0.0006010052748024464, 0.05679146572947502, 0.05045616254210472, -0.021494777873158455, 0.00848478078842163, -0.014480692334473133, 0.0025595754850655794, 0.008824171498417854, -0.0000329227150359656...
[ "F", "False", "Model", "ModelIntegral", "ModelLQE", "ModelLayer", "ModelMLP", "ModelMLPPredictionHead", "ModelOutput", "ModelPreTrainedModel", "ModuleList", "None", "Parameter", "_", "__init__", "all_cross_attentions", "all_hidden_states", "all_self_attns", "and", "attentions",...
d_fine/modeling_d_fine.py:DFineFrozenBatchNorm2d
[ -0.0003010575601365417, 0.02442867122590542, 0.008888366632163525, 0.012501057237386703, -0.0010680375853553414, 0.03922349959611893, -0.00992056354880333, -0.022134898230433464, 0.009289775975048542, 0.015827026218175888, 0.014336074702441692, 0.0027238542679697275, 0.0008816685876809061, ...
[ "Model", "Module", "__init__", "_load_from_state_dict", "bias", "class", "def", "del", "epsilon", "error_msgs", "forward", "if", "in", "local_metadata", "missing_keys", "n", "nn", "num_batches_tracked", "num_batches_tracked_key", "ones", "prefix", "register_buffer", "resh...
d_fine/modeling_d_fine.py:DFineModelOutput
[ -0.0001226391614181921, 0.02445673756301403, 0.007905783131718636, 0.01865537278354168, -0.0007571635651402175, 0.06597631424665451, 0.0646112933754921, -0.021157922223210335, 0.014674043282866478, -0.006995764560997486, 0.000604308967012912, 0.013877777382731438, -0.003441006178036332, 0....
[ "Model", "ModelOutput", "None", "class", "cross_attentions", "decoder_attentions", "decoder_hidden_states", "denoising_meta_values", "enc_outputs_class", "enc_outputs_coord_logits", "enc_topk_bboxes", "enc_topk_logits", "encoder_attentions", "encoder_hidden_states", "encoder_last_hidden_...
d_fine/modeling_d_fine.py:replace_batch_norm
[ 0.00006147739622974768, 0.005075003486126661, 0.02018596976995468, -0.013628380373120308, 0.0004401425248943269, 0.03170451521873474, -0.014312650077044964, -0.013457313179969788, 0.028169121593236923, 0.00610140897333622, 0.011119389906525612, -0.007698039058595896, 0.0019815322011709213, ...
[ "BatchNorm2d", "ModelFrozenBatchNorm2d", "Model_batch_norm", "_modules", "bias", "children", "copy_", "def", "device", "for", "if", "in", "isinstance", "len", "list", "meta", "model", "module", "name", "named_children", "new_module", "nn", "num_features", "r", "runnin...
d_fine/modeling_d_fine.py:DFineConvEncoder
[ -0.00022557828924618661, 0.009946758858859539, 0.031829629093408585, 0.011822434142231941, -0.0007708738557994366, 0.02591841295361519, 0.01716526411473751, -0.04069645702838898, 0.006820634938776493, -0.0048028635792434216, 0.020461905747652054, 0.0011225627968087792, -0.00046181381912902, ...
[ "Model", "Module", "None", "__init__", "append", "backbone", "bool", "channels", "class", "config", "def", "feature_map", "feature_maps", "features", "float", "for", "forward", "freeze_backbone_batch_norms", "functional", "if", "in", "intermediate_channel_sizes", "interpo...
d_fine/modeling_d_fine.py:get_contrastive_denoising_training_group
[ -0.00012422721192706376, 0.0352095402777195, -0.010278914123773575, -0.0026691104285418987, -0.00021118625591043383, 0.05179210007190704, 0.0027117026038467884, -0.02510099485516548, 0.01487887091934681, -0.0008766892133280635, 0.008575227111577988, 0.007609804160892963, -0.00021739762451034...
[ "Model_contrastive_denoising_training_group", "None", "attn_mask", "batch_size", "bool", "box_noise_scale", "boxes", "center_to_corners_format", "class_embed", "class_labels", "clip_", "corners_to_center_format", "def", "denoise_positive_idx", "denoising_meta_values", "device", "diff...
d_fine/modeling_d_fine.py:DFineModel
[ -0.00018712809833232313, 0.02734188735485077, -0.0029799267649650574, 0.019433077424764633, -0.0007485123933292925, 0.03977001830935478, 0.03163523972034454, -0.012258656322956085, 0.007033192086964846, -0.008586708456277847, -0.0058186249807477, 0.015026739798486233, -0.003347121411934495, ...
[ "BaseModelOutput", "BatchNorm2d", "Conv2d", "Embedding", "F", "False", "Identity", "LayerNorm", "Linear", "Model", "ModelConvEncoder", "ModelDecoder", "ModelHybridEncoder", "ModelMLPPredictionHead", "ModelOutput", "ModelPreTrainedModel", "ModuleList", "None", "Sequential", "Tru...
d_fine/modeling_d_fine.py:DFineObjectDetectionOutput
[ -0.00010273321095155552, 0.017911003902554512, 0.0060931420885026455, 0.01598387025296688, -0.0006801647250540555, 0.0675630271434784, 0.06302859634160995, -0.007481812033802271, 0.014963624067604542, -0.013376573100686073, -0.006433224771171808, 0.007481812033802271, -0.0035992050543427467,...
[ "Model", "ModelOutput", "None", "auxiliary_outputs", "class", "cross_attentions", "decoder_attentions", "decoder_hidden_states", "denoising_meta_values", "enc_outputs_class", "enc_outputs_coord_logits", "enc_topk_bboxes", "enc_topk_logits", "encoder_attentions", "encoder_hidden_states", ...
d_fine/modeling_d_fine.py:DFineForObjectDetection
[ -0.00013067632971797138, 0.026605350896716118, -0.01044007483869791, 0.01100136898458004, -0.0004069384012836963, 0.04063770920038223, 0.03929060325026512, -0.009317485615611076, 0.00645488454028964, -0.0033677660394459963, 0.010720721445977688, 0.037494461983442307, -0.0014102519489824772, ...
[ "Linear", "Model", "ModelMLP", "ModelModel", "ModelObjectDetectionOutput", "ModelPreTrainedModel", "ModuleList", "None", "_", "__init__", "_no_split_modules", "_set_aux_loss", "_tied_weights_keys", "a", "auto_docstring", "auxiliary_outputs", "b", "bbox_embed", "class", "class_e...
d_fine/modeling_d_fine.py:DFineMLPPredictionHead
[ 0, 0.008214556612074375, 0.01801537163555622, 0.045774906873703, 0.00045675766887143254, 0.050080470740795135, 0.017108937725424767, -0.006968209985643625, 0.005778515711426735, 0.0072514708153903484, 0.01801537163555622, -0.06662288308143616, 0.00048154298565350473, 0.022547541186213493, ...
[ "Linear", "Model", "Module", "ModuleList", "__init__", "class", "config", "d_model", "def", "else", "enumerate", "for", "forward", "functional", "h", "i", "if", "in", "input_dim", "k", "layer", "layers", "n", "nn", "num_layers", "output_dim", "relu", "return", ...
d_fine/modeling_d_fine.py:DFineMLP
[ -0.0001238625409314409, 0.004388272762298584, 0.015061684884130955, 0.027745207771658897, -0.00008847324352245778, 0.0536784864962101, 0.02593327686190605, -0.01291001494973898, 0.008946414105594158, 0.002930233720690012, 0.029670385643839836, -0.06024673953652382, -0.0003202731313649565, ...
[ "ACT2CLS", "Linear", "Model", "Module", "ModuleList", "__init__", "act", "class", "def", "else", "enumerate", "for", "forward", "hidden_dim", "hidden_dims", "i", "if", "in", "in_dim", "input_dim", "input_dims", "layer", "layers", "nn", "num_layers", "out_dim", "ou...
d_fine/modeling_d_fine.py:DFineLQE
[ -0.0003461897140368819, 0.033408213406801224, 0.017516111955046654, 0.0096860621124506, -0.0013847588561475277, 0.03851224482059479, 0.060784388333559036, -0.016588104888796806, 0.009338059462606907, 0.008468054234981537, 0.045472290366888046, 0.0034365218598395586, 0.0004893781151622534, ...
[ "F", "Model", "ModelMLP", "Module", "True", "_", "__init__", "batch_size", "cat", "class", "config", "def", "dim", "forward", "keepdim", "length", "lqe_hidden_dim", "lqe_layers", "max_num_bins", "mean", "nn", "pred_corners", "prob", "prob_topk", "quality_score", "re...
d_fine/modeling_d_fine.py:DFineConvNormLayer
[ -0.00011776775500038639, 0.009123901836574078, 0.004788631573319435, -0.001154655241407454, -0.00027449626941233873, 0.02686167322099209, 0.012920805253088474, -0.04556284099817276, 0.015414293855428696, -0.004618620965629816, 0.048509690910577774, -0.028561780229210854, 0.001657603308558464...
[ "ACT2CLS", "BatchNorm2d", "Conv2d", "Identity", "Model", "Module", "None", "__init__", "activation", "batch_norm_eps", "class", "config", "conv", "def", "else", "forward", "groups", "hidden_state", "if", "in_channels", "is", "kernel_size", "nn", "norm", "out_channels"...
d_fine/modeling_d_fine.py:DFineRepVggBlock
[ 0.00006704460247419775, 0.02011870965361595, 0.001903889118693769, 0.008127049542963505, 0.0006677819765172899, 0.03500882536172867, 0.007445058785378933, -0.024892639368772507, 0.018072739243507385, -0.005882164929062128, 0.05001261085271835, -0.023528659716248512, 0.00291266618296504, -0...
[ "ACT2CLS", "Identity", "Model", "ModelConvNormLayer", "Module", "None", "__init__", "activation", "activation_function", "class", "config", "conv1", "conv2", "def", "else", "forward", "hidden_channels", "if", "in_channels", "is", "nn", "out_channels", "padding", "return...
d_fine/modeling_d_fine.py:DFineCSPRepLayer
[ 0, -0.0024119848385453224, 0.0018125567585229874, -0.001512842602096498, 0.0005851560854353011, 0.06348229944705963, -0.010104646906256676, -0.03585151582956314, 0.008220729418098927, -0.007278771139681339, 0.046355780214071274, -0.05526157096028328, 0.002626066328957677, 0.032197859138250...
[ "Identity", "Model", "ModelConvNormLayer", "ModelRepVggBlock", "Module", "ModuleList", "_", "__init__", "activation", "activation_function", "bottleneck", "bottlenecks", "class", "config", "conv1", "conv2", "conv3", "def", "else", "expansion", "for", "forward", "hidden_ch...
d_fine/modeling_d_fine.py:DFineRepNCSPELAN4
[ -0.0002079914411297068, 0.02155187539756298, 0.005676095373928547, 0.008413298986852169, -0.0006518865702673793, 0.052323803305625916, 0.01728760078549385, -0.052093300968408585, 0.010314934886991978, -0.009047177620232105, 0.020399367436766624, -0.041720740497112274, 0.002175356261432171, ...
[ "Model", "ModelCSPRepLayer", "ModelConvNormLayer", "Module", "__init__", "act", "activation", "branch1", "branch2", "cat", "class", "config", "conv1", "conv1_dim", "conv2", "conv2_dim", "conv3", "conv3_dim", "conv4", "conv4_dim", "conv_dim", "csp_rep1", "csp_rep2", "def...
d_fine/modeling_d_fine.py:DFineSCDown
[ 0.0000874908801051788, 0.030469920486211777, 0.011028291657567024, -0.008754417300224304, 0.0008384912507608533, 0.03092469461262226, 0.004263515118509531, -0.02569478377699852, 0.020237484946846962, 0.005883650854229927, 0.0441131666302681, -0.02330721542239189, 0.004860407207161188, -0.0...
[ "Model", "ModelConvNormLayer", "Module", "__init__", "class", "config", "conv1", "conv2", "def", "encoder_hidden_dim", "forward", "input_features", "kernel_size", "nn", "return", "self", "stride", "super" ]
d_fine/modeling_d_fine.py:DFineEncoderLayer
[ -0.00019330305804032832, 0.07479847967624664, 0.022006260231137276, 0.018471576273441315, -0.0009264293476007879, 0.02839149720966816, 0.026339098811149597, -0.007468446157872677, 0.007753501180559397, 0.014708847738802433, 0.02497083507478237, 0.01596309058368206, 0.003591695334762335, -0...
[ "ACT2FN", "False", "LayerNorm", "Linear", "Model", "ModelMultiheadAttention", "Module", "None", "__init__", "activation_dropout", "activation_fn", "any", "attention_mask", "attn_weights", "clamp", "clamp_value", "class", "config", "def", "dropout", "dtype", "embed_dim", "...
d_fine/modeling_d_fine.py:DFineEncoder
[ -0.00006442629091907293, 0.02937132865190506, 0.01547643169760704, 0.0325343944132328, -0.00027535620029084384, 0.048575662076473236, 0.02677309699356556, -0.014120831154286861, 0.0116920480504632, -0.021011797711253166, 0, -0.012595782056450844, 0.00045539680286310613, -0.0030783412512391...
[ "False", "Model", "ModelLayer", "Module", "ModuleList", "None", "_", "__init__", "attention_mask", "class", "config", "def", "encoder_layers", "for", "forward", "hidden_states", "in", "layer", "layers", "nn", "output_attentions", "pos_embed", "position_embeddings", "ran...
d_fine/modeling_d_fine.py:DFineHybridEncoder
[ -0.00016133865574374795, 0.028026828542351723, -0.006439362186938524, 0.04152963310480118, -0.000421962613472715, 0.028253765776753426, 0.0304096769541502, -0.026665201410651207, 0.004453655797988176, -0.0047373282723128796, 0.010836283676326275, 0.005616712383925915, 0.001531830639578402, ...
[ "BaseModelOutput", "Embed", "F", "Model", "ModelConvNormLayer", "ModelEncoder", "ModelRepNCSPELAN4", "ModelSCDown", "Module", "ModuleList", "None", "ValueError", "_", "__init__", "all_attentions", "append", "arange", "attention_mask", "attentions", "backbone_feature_map", "be...
squeezebert/modeling_squeezebert.py:SqueezeBertEmbeddings
[ -0.00018897172412835062, 0.017374752089381218, 0.014195059426128864, -0.0018453577067703009, -0.0010504344245418906, 0.041790254414081573, 0.032478295266628265, 0.0029099872335791588, 0.008289914578199387, -0.010390783660113811, 0.02452906221151352, 0.024415502324700356, -0.00042407738510519...
[ "Dropout", "Embedding", "False", "LayerNorm", "ModelEmbeddings", "Module", "None", "__init__", "arange", "class", "config", "def", "device", "dropout", "dtype", "else", "embedding_size", "embeddings", "eps", "expand", "forward", "hidden_dropout_prob", "hidden_size", "if...
squeezebert/modeling_squeezebert.py:MatMulWrapper
[ 0.00004202286072541028, 0.001380119239911437, 0.0071341549046337605, 0.0021232604049146175, 0.0007183697307482362, 0.06069693714380264, 0.004218210466206074, -0.013079283758997917, 0.009455585852265358, 0.013022663071751595, 0.01596691831946373, -0.04552270099520683, -0.0029442543163895607, ...
[ "Model1", "Model2", "ModelMulWrapper", "Modelmul", "Module", "__init__", "class", "def", "forward", "nn", "return", "self", "super", "torch" ]
squeezebert/modeling_squeezebert.py:SqueezeBertLayerNorm
[ 0.000021813502826262265, 0.03434738889336586, 0.03524535894393921, 0.04669448733329773, 0.00009339248208561912, 0.03300043195486069, 0.00976543314754963, -0.0249186921864748, 0.015602244064211845, 0.04467405378818512, 0.0033814217895269394, 0.004742408636957407, 0.0007436321466229856, -0.0...
[ "LayerNorm", "ModelLayerNorm", "__init__", "class", "def", "eps", "forward", "hidden_size", "nn", "normalized_shape", "permute", "return", "self", "x" ]
squeezebert/modeling_squeezebert.py:ConvDropoutLayerNorm
[ -0.00005248138768365607, 0.03890034556388855, 0.017763741314411163, 0.00792622659355402, 0, 0.04744493216276169, 0.008938085287809372, -0.026870470494031906, 0.006127366796135902, 0.005537115503102541, 0.02597104012966156, -0.00041633768705651164, 0.002192360581830144, -0.02259817719459533...
[ "Dropout", "Model1d", "ModelDropoutLayerNorm", "ModelLayerNorm", "Module", "__init__", "cin", "class", "cout", "def", "dropout", "dropout_prob", "forward", "groups", "hidden_states", "in_channels", "input_tensor", "kernel_size", "layernorm", "nn", "out_channels", "return", ...
squeezebert/modeling_squeezebert.py:ConvActivation
[ 0.0000639461723039858, 0.009031846188008785, -0.00999148003757, 0.01145915500819683, 0.0003651547012850642, 0.054416872560977936, 0.016708916053175926, -0.018176591023802757, 0.01089466456323862, -0.028789009898900986, 0.03951432555913925, -0.03815954923629761, 0.002554318867623806, -0.011...
[ "ACT2FN", "Model1d", "ModelActivation", "Module", "__init__", "act", "cin", "class", "cout", "def", "forward", "groups", "in_channels", "kernel_size", "nn", "out_channels", "output", "return", "self", "super", "x" ]
squeezebert/modeling_squeezebert.py:SqueezeBertSelfAttention
[ 0.00010328183270758018, 0.0375787653028965, 0.018564360216259956, -0.011194871738553047, 0.0003920314775314182, 0.009563458152115345, 0.006947571411728859, -0.006750676780939102, 0.00787578895688057, 0.02385239116847515, 0.024077413603663445, 0.010013503953814507, 0.002250225516036153, -0....
[ "Conv1d", "Dropout", "MatMulWrapper", "ModelSelfAttention", "Module", "Softmax", "ValueError", "__init__", "a", "all_head_size", "attention", "attention_head_size", "attention_mask", "attention_probs", "attention_probs_dropout_prob", "attention_score", "cin", "class", "config", ...
squeezebert/modeling_squeezebert.py:SqueezeBertModule
[ -0.00012300240632612258, 0.026278268545866013, 0.009004837833344936, 0.004360833205282688, -0.00047785104834474623, 0.04326852783560753, 0.022087337449193, -0.020501580089330673, 0.008042056113481522, -0.005465200170874596, 0.0342070572078228, -0.0037661741953343153, 0.003737857099622488, ...
[ "ConvActivation", "ConvDropoutLayerNorm", "ModelModule", "ModelSelfAttention", "Module", "__init__", "act", "att", "attention", "attention_mask", "attention_output", "attention_score", "c0", "c1", "c2", "c3", "cin", "class", "config", "context_layer", "cout", "def", "drop...
squeezebert/modeling_squeezebert.py:SqueezeBertEncoder
[ -0.00001174934186565224, 0.020464498549699783, 0.021364036947488785, 0.02125159464776516, 0.00010058315092464909, 0.034632228314876556, 0.010288470424711704, -0.03845526650547981, 0.012537316419184208, 0.009951143525540829, 0.01821565255522728, 0.0018693532329052687, -0.0012228100094944239, ...
[ "BaseModelOutput", "Conv1d", "False", "If", "ModelEncoder", "ModelModule", "Module", "ModuleList", "None", "True", "_", "__init__", "a", "adjust", "all_attentions", "all_hidden_states", "assert", "attention_mask", "attention_score", "attentions", "before", "channels", "cl...
squeezebert/modeling_squeezebert.py:SqueezeBertPooler
[ -0.0002669636451173574, 0.014238061383366585, 0.03644943609833717, 0.021755756810307503, -0.0010464974911883473, 0.030526403337717056, 0.035538200289011, -0.003644943702965975, -0.0033459444530308247, 0.004897892940789461, 0.014807583764195442, -0.01389634795486927, -0.001409568008966744, ...
[ "Linear", "ModelPooler", "Module", "Tanh", "__init__", "activation", "class", "config", "def", "dense", "first_token_tensor", "forward", "hidden_size", "hidden_states", "nn", "pooled_output", "return", "self", "super" ]
squeezebert/modeling_squeezebert.py:SqueezeBertPredictionHeadTransform
[ -0.0003230067668482661, 0.04226364195346832, 0.04410118982195854, 0.03054926171898842, -0.0013781621819362044, 0.026874162256717682, 0.029860179871320724, -0.01056591048836708, -0.0034741172567009926, 0.02193574793636799, 0.01527463085949421, 0.007694738917052746, -0.00010632305929902941, ...
[ "ACT2FN", "LayerNorm", "Linear", "ModelPredictionHeadTransform", "Module", "__init__", "class", "config", "def", "dense", "else", "eps", "forward", "hidden_act", "hidden_size", "hidden_states", "if", "isinstance", "layer_norm_eps", "nn", "return", "self", "str", "super"...
squeezebert/modeling_squeezebert.py:SqueezeBertLMPredictionHead
[ -0.00035434559686109424, 0.03384450078010559, 0.03200262412428856, 0.01922459714114666, -0.0016044480726122856, 0.022793235257267952, 0.05433538928627968, -0.0275130458176136, -0.0006655221804976463, 0.004633473232388496, 0.019685067236423492, 0.006763143930584192, 0.000031252562621375546, ...
[ "Linear", "ModelLMPredictionHead", "ModelPredictionHeadTransform", "Module", "Parameter", "__init__", "bias", "class", "config", "decoder", "def", "forward", "hidden_size", "hidden_states", "nn", "return", "self", "super", "torch", "transform", "vocab_size", "zeros" ]
squeezebert/modeling_squeezebert.py:SqueezeBertOnlyMLMHead
[ -0.00035440587089397013, 0.019890347495675087, 0.019774029031395912, 0.03117317706346512, -0.0017011481104418635, 0.040478602051734924, 0.04373550042510033, -0.00942174345254898, -0.0026753100100904703, -0.004681792575865984, 0.016400812193751335, 0.010526763275265694, -0.0022100387141108513...
[ "ModelLMPredictionHead", "ModelOnlyMLMHead", "Module", "__init__", "class", "config", "def", "forward", "nn", "prediction_scores", "predictions", "return", "self", "sequence_output", "super" ]
squeezebert/modeling_squeezebert.py:SqueezeBertPreTrainedModel
[ -0.0002844761766027659, 0.03755801171064377, -0.0006047355709597468, -0.00967576913535595, -0.0014527966268360615, 0.03091665729880333, 0.024847831577062607, -0.008874225430190563, 0.0006834585219621658, -0.006183331366628408, 0.0023760024923831224, 0.006727235857397318, -0.00086237437790259...
[ "ModelConfig", "ModelEmbeddings", "ModelLMPredictionHead", "ModelPreTrainedModel", "PreTrainedModel", "_init_weights", "arange", "base_model_prefix", "bias", "class", "config", "copy_", "def", "elif", "expand", "if", "init", "isinstance", "module", "no_grad", "position_ids", ...
squeezebert/modeling_squeezebert.py:SqueezeBertModel
[ -0.000010670204574125819, 0.01938715949654579, 0.011878837831318378, 0.018490644171833992, -0.00009149016841547564, 0.03406761214137077, 0.014568386599421501, 0.0014008063590154052, 0.006443709135055542, 0.0045386129058897495, 0.020619869232177734, 0.0014778507174924016, -0.00045176004641689...
[ "BaseModelOutputWithPooling", "ModelEmbeddings", "ModelEncoder", "ModelModel", "ModelPooler", "ModelPreTrainedModel", "None", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "class", "config", "def", "device", "dtype", "elif", "else", "embedding_output", ...
squeezebert/modeling_squeezebert.py:SqueezeBertForMaskedLM
[ -0.00011411308514652774, 0.04103193059563637, 0.018954968079924583, -0.012376478873193264, -0.0005714365397579968, 0.039916932582855225, 0.01549847424030304, -0.008808485232293606, 0.0035122439730912447, -0.009198734536767006, 0.01806296966969967, 0.04036293178796768, -0.0006829363410361111,...
[ "CrossEntropyLoss", "MaskedLMOutput", "ModelForMaskedLM", "ModelModel", "ModelOnlyMLMHead", "ModelPreTrainedModel", "None", "__init__", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "bias", "class", "cls", "config", "decoder", "def", "else", "embedding...
squeezebert/modeling_squeezebert.py:SqueezeBertForSequenceClassification
[ -0.00036776973865926266, 0.03513367101550102, 0.010620472021400928, 0.009242681786417961, -0.0012988550588488579, 0.028703980147838593, 0.0202076006680727, 0.009185273200273514, -0.0058843158185482025, 0.007750074379146099, 0.05419311299920082, -0.00021348585141822696, 0.00030497979605570436...
[ "BCEWithLogitsLoss", "CrossEntropyLoss", "Dropout", "Linear", "MSELoss", "Model", "ModelForSequenceClassification", "ModelModel", "ModelPreTrainedModel", "None", "SequenceClassifierOutput", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "class", "classifier", ...
squeezebert/modeling_squeezebert.py:SqueezeBertForMultipleChoice
[ -0.00026702144532464445, 0.05878744274377823, 0.02722906693816185, 0.030305154621601105, -0.0009968801168724895, 0.04055877774953842, 0.04579951986670494, -0.003047604812309146, -0.0016377315623685718, 0.008487721905112267, 0.03622946888208389, -0.006209138780832291, -0.0012104972265660763, ...
[ "CrossEntropyLoss", "Dropout", "Linear", "ModelForMultipleChoice", "ModelModel", "ModelPreTrainedModel", "MultipleChoiceModelOutput", "None", "__init__", "attention_mask", "attentions", "auto_docstring", "class", "classifier", "config", "def", "dropout", "else", "forward", "hid...
squeezebert/modeling_squeezebert.py:SqueezeBertForTokenClassification
[ -0.00027394964126870036, 0.03734253719449043, 0.0056070731952786446, -0.03165007755160332, -0.0009890649234876037, 0.03779793530702591, 0.04053031653165817, 0.015141943469643593, -0.0021062102168798447, 0.007343273144215345, 0.05441991612315178, 0.01958206295967102, -0.0007400197791866958, ...
[ "CrossEntropyLoss", "Dropout", "Linear", "ModelForTokenClassification", "ModelModel", "ModelPreTrainedModel", "None", "TokenClassifierOutput", "__init__", "attention_mask", "attentions", "auto_docstring", "class", "classifier", "config", "def", "dropout", "else", "forward", "hi...
squeezebert/modeling_squeezebert.py:SqueezeBertForQuestionAnswering
[ -0.00020632563973776996, 0.031601328402757645, 0.02031514048576355, 0.012019791640341282, -0.0007865062798373401, 0.047401994466781616, 0.03950166329741478, 0.02550678700208664, 0.0032588872127234936, 0.027086853981018066, 0.019976554438471794, 0.020428001880645752, 0.0010862956987693906, ...
[ "CrossEntropyLoss", "Linear", "Model", "ModelForQuestionAnswering", "ModelModel", "ModelPreTrainedModel", "None", "QuestionAnsweringModelOutput", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "clamp", "class", "config", "contiguous", "def", "dim", "else",...
bit/modeling_bit.py:get_padding_value
[ 0, 0.005243157967925072, 0.012572303414344788, -0.029767606407403946, 0.00048273697029799223, 0.010993718169629574, 0.014150888659060001, -0.04668101668357849, 0.016349632292985916, 0.0006906309863552451, 0.03901360556483269, -0.036983996629714966, 0.0049894568510353565, -0.019168533384799...
[ "False", "Model_padding_value", "None", "True", "and", "def", "dilation", "dynamic", "elif", "else", "if", "is", "isinstance", "kernel_size", "lower", "padding", "r", "return", "same", "str", "stride", "valid" ]
bit/modeling_bit.py:WeightStandardizedConv2d
[ -0.00008420228550676256, 0.006799665279686451, 0.020652925595641136, -0.012301469221711159, 0.00017986666352953762, 0.02076578140258789, -0.003145903116092086, -0.04898016154766083, 0.017380056902766228, 0.021217212080955505, 0.03408296778798103, -0.01568719372153282, 0.0032164391595870256, ...
[ "Conv2d", "DynamicPad2d", "Model", "ModelStandardizedConv2d", "None", "SAME", "True", "__init__", "batch_norm", "bias", "class", "conv2d", "def", "dilation", "else", "eps", "forward", "functional", "get_padding_value", "groups", "hidden_state", "if", "in_channel", "is",...
bit/modeling_bit.py:BitGroupNormActivation
[ -0.00009721535752760246, 0.014140415005385876, 0.0016120073851197958, 0.022398417815566063, -0.0002757381007540971, 0.027488967403769493, 0.008427687920629978, -0.017647238448262215, 0.006504591088742018, -0.005882412660866976, 0.0452493280172348, 0.0010746715124696493, -0.000121961078548338...
[ "ACT2FN", "GroupNorm", "Identity", "ModelGroupNormActivation", "True", "__init__", "activation", "affine", "apply_activation", "bias", "class", "config", "def", "else", "eps", "forward", "functional", "group_norm", "hidden_act", "hidden_state", "if", "nn", "num_channels",...
bit/modeling_bit.py:DynamicPad2d
[ -0.00011321489000692964, -0.013480470515787601, 0.0002404719270998612, -0.03976738825440407, -0.00009653982851887122, 0.02033304236829281, 0.008481462486088276, -0.03437519818544388, 0.010672039352357388, 0.019209669902920723, 0.019097333773970604, -0.047855667769908905, 0.003987972624599934...
[ "ModelPad2d", "Module", "__init__", "ceil", "class", "compute_padding", "def", "dilation", "forward", "functional", "if", "input", "input_height", "input_width", "int", "isinstance", "kernel_size", "math", "max", "nn", "or", "pad", "padding_height", "padding_width", "...
bit/modeling_bit.py:BitMaxPool2d
[ -0.0001517121709184721, -0.014816929586231709, 0.030531853437423706, -0.01251581497490406, -0.00018415927479509264, 0.016949668526649475, 0.0059492215514183044, -0.05073675885796547, 0.011056572198867798, 0.024021385237574577, 0.03816481679677963, -0.04557328298687935, 0.0026098000817000866,...
[ "DynamicPad2d", "False", "Identity", "Iterable", "MaxPool2d", "ModelMaxPool2d", "None", "True", "__init__", "abc", "ceil_mode", "class", "collections", "def", "dilation", "else", "forward", "functional", "hidden_states", "if", "isinstance", "kernel_size", "max_pool2d", ...
bit/modeling_bit.py:BitEmbeddings
[ -0.00003605594247346744, 0.014070611447095871, 0.020261680707335472, 0.02240041270852089, 0.00010289134661434218, 0.022625543177127838, -0.00013806787319481373, -0.026002489030361176, 0.011143924668431282, 0.004558878019452095, 0.03669615462422371, 0.005487538408488035, 0.002870404627174139,...
[ "ConstantPad2d", "Identity", "Make", "ModelEmbeddings", "ModelGroupNormActivation", "ModelMaxPool2d", "Module", "None", "SAME", "ValueError", "WeightStandardizedConv2d", "__init__", "and", "channel", "class", "config", "configuration", "convolution", "def", "dimension", "else...
bit/modeling_bit.py:drop_path
[ 0, 0.017753854393959045, 0.03775503486394882, -0.02123720571398735, 0, 0.043598074465990067, 0.0339345820248127, -0.027304979041218758, 0.014832334592938423, 0.0017486985307186842, 0.020113544538617134, -0.04944111406803131, 0.0016574009787291288, -0.018315685912966728, -0.05034004524350...
[ "False", "Model_path", "Model_prob", "def", "device", "div", "dtype", "floor_", "if", "input", "keep_prob", "ndim", "not", "or", "output", "rand", "random_tensor", "return", "shape", "torch", "training" ]
bit/modeling_bit.py:BitDropPath
[ -0.00011107319733127952, 0.018094833940267563, 0.03596488758921623, -0.025063030421733856, -0.00047414645086973906, 0.031918834894895554, 0.033717080950737, -0.026861274614930153, 0.0066872211173176765, -0.002430439693853259, 0.009721758775413036, -0.033717080950737, -0.000818341679405421, ...
[ "ModelDropPath", "Module", "None", "__init__", "class", "def", "drop_path", "drop_prob", "extra_repr", "f", "forward", "hidden_states", "nn", "p", "return", "self", "super", "training" ]
bit/modeling_bit.py:make_div
[ -0.00010597718937788159, 0.02555413544178009, 0.010356741957366467, -0.018687164410948753, -0.0001811726251617074, 0.04165211319923401, 0.04592989757657051, -0.022176936268806458, 0.008949575945734978, 0.03850006312131882, 0.0013438435271382332, -0.0533597357571125, 0.0018433874938637018, ...
[ "Model_div", "def", "divisor", "if", "int", "max", "min_value", "new_value", "return", "value" ]
bit/modeling_bit.py:BitPreActivationBottleneckLayer
[ -0.00032858102349564433, 0.0028584753163158894, 0.009882568381726742, -0.02160375379025936, -0.001019858056679368, 0.05286024883389473, -0.0031170309521257877, -0.02895822376012802, 0.00045606328058056533, -0.017926519736647606, 0.02769417315721512, -0.03010735847055912, 0.001515422831289470...
[ "False", "Identity", "ModelDownsampleConv", "ModelDropPath", "ModelGroupNormActivation", "ModelPreActivationBottleneckLayer", "Module", "None", "True", "WeightStandardizedConv2d", "__init__", "bottle_ratio", "class", "config", "conv1", "conv2", "conv3", "def", "dilation", "down...
bit/modeling_bit.py:BitBottleneckLayer
[ -0.0002280639309901744, -0.003677530912682414, 0.01482415571808815, -0.04013925418257713, -0.0005844138213433325, 0.05405115336179733, 0.009179573506116867, -0.04013925418257713, 0.005502042360603809, -0.007982238195836544, 0.02668347954750061, -0.03056056797504425, 0.0027082592714577913, ...
[ "ACT2FN", "False", "Identity", "ModelBottleneckLayer", "ModelDownsampleConv", "ModelDropPath", "ModelGroupNormActivation", "Module", "None", "WeightStandardizedConv2d", "__init__", "activation", "apply_activation", "bottle_ratio", "class", "config", "conv1", "conv2", "conv3", "...
bit/modeling_bit.py:BitDownsampleConv
[ -0.00011145230382680893, 0.03100304864346981, 0.021455008536577225, 0.013479585759341717, -0.0001904342498164624, 0.026172863319516182, 0.017860451713204384, -0.029879748821258545, 0.009660369716584682, 0.01673715189099312, 0.01297410111874342, 0.0013268967159092426, 0.0035383913200348616, ...
[ "False", "Identity", "ModelDownsampleConv", "ModelGroupNormActivation", "Module", "True", "WeightStandardizedConv2d", "__init__", "apply_activation", "class", "config", "conv", "def", "else", "eps", "forward", "global_padding", "if", "in_channels", "nn", "norm", "num_channe...
bit/modeling_bit.py:BitStage
[ -0.0001959261135198176, -0.024508576840162277, -0.010886367410421371, -0.03032223880290985, -0.0001959261135198176, 0.05562876909971237, 0.0026503459084779024, -0.02336864173412323, -0.005300691816955805, -0.0006518996087834239, 0.04742124676704407, -0.05813662335276604, 0.000162084339535795...
[ "ModelBottleneckLayer", "ModelPreActivationBottleneckLayer", "ModelStage", "Module", "None", "Sequential", "_", "__init__", "_get_updated_hyperparameters", "add_module", "bottle_ratio", "bottleneck", "class", "config", "def", "depth", "dilation", "drop_path_rate", "else", "enum...
bit/modeling_bit.py:BitEncoder
[ -0.00027007426251657307, 0.010473872534930706, 0.01293494738638401, -0.0029904909897595644, -0.0006975427968427539, 0.042582303285598755, 0.011618558317422867, -0.037545688450336456, -0.0014093940844759345, 0.005523107945919037, 0.03159332275390625, -0.020718809217214584, -0.0011232227552682...
[ "BaseModelOutputWithNoAttention", "False", "ModelEncoder", "ModelStage", "Module", "ModuleList", "None", "Tensor", "True", "__init__", "_get_updated_hyperparameters", "add_module", "class", "config", "current_depth", "current_hidden_size", "current_stride", "def", "depth", "dep...
bit/modeling_bit.py:BitPreTrainedModel
[ -0.00004603287743520923, 0.04130157455801964, -0.00707064988091588, 0.017957204952836037, -0.0001972837490029633, 0.023344367742538452, 0.010437625460326672, -0.009034719318151474, 0.007856277748942375, 0.01941622793674469, 0.01251392811536789, -0.01941622793674469, -0.0013467904645949602, ...
[ "BatchNorm2d", "Conv2d", "GroupNorm", "Linear", "Model", "ModelConfig", "ModelEmbeddings", "ModelPreTrainedModel", "None", "PreTrainedModel", "_", "_calculate_fan_in_and_fan_out", "_init_weights", "_no_split_modules", "a", "base_model_prefix", "bias", "bound", "class", "config"...
bit/modeling_bit.py:BitModel
[ 0.000055535499996040016, 0.03214298188686371, -0.004270745906978846, 0.0418083555996418, 0.0003810657362919301, 0.03776238486170769, 0.012756044045090675, -0.015284774824976921, 0.01163216307759285, 0.023826265707612038, 0.03528984636068344, -0.003961678594350815, 0.002332051983103156, -0....
[ "AdaptiveAvgPool2d", "BaseModelOutputWithPoolingAndNoAttention", "Identity", "ModelEmbeddings", "ModelEncoder", "ModelGroupNormActivation", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "auto_docstring", "class", "config", "def", "else", "embedder", "embedding_output", ...
bit/modeling_bit.py:BitForImageClassification
[ -0.000042190989915980026, 0.031177837401628494, 0.008740929886698723, 0.027837354689836502, 0.00009003644663607702, 0.006207730155438185, 0.02171313762664795, -0.008295531384646893, 0.0046488381922245026, 0.00823985692113638, 0.047212153673172, 0.009631725028157234, 0.000953429436776787, -...
[ "Flatten", "Identity", "ImageClassifierOutputWithNoAttention", "Linear", "Model", "ModelForImageClassification", "ModelModel", "ModelPreTrainedModel", "None", "Sequential", "__init__", "auto_docstring", "class", "classifier", "config", "def", "else", "forward", "hidden_sizes", ...
bit/modeling_bit.py:BitBackbone
[ -0.0001899114140542224, -0.0073374067433178425, 0.03271027281880379, 0.02408461831510067, -0.0005741100758314133, 0.021508123725652695, 0.0037247142754495144, -0.019043652340769768, -0.0009031731751747429, 0.0015683007659390569, -0.004060778766870499, 0.002212424296885729, -0.000630120804999...
[ "BackboneMixin", "BackboneOutput", "False", "Model", "ModelBackbone", "ModelModel", "ModelPreTrainedModel", "None", "True", "__init__", "_init_backbone", "attentions", "auto_docstring", "class", "config", "def", "else", "embedding_size", "enumerate", "feature_maps", "for", ...
deberta/modeling_deberta.py:DebertaLayerNorm
[ -0.000020909797967760824, 0.03071616403758526, 0.03407924622297287, 0.04618635028600693, 0.00008933190838433802, 0.03654550760984421, 0.017936445772647858, -0.008912171237170696, 0.013844694010913372, 0.03475186601281166, 0.022981071844697, 0.003503212006762624, 0.003180916653946042, 0.005...
[ "ModelLayerNorm", "Module", "Parameter", "True", "__init__", "bias", "class", "def", "eps", "float", "forward", "hidden_states", "keepdim", "mean", "nn", "ones", "pow", "return", "self", "size", "sqrt", "super", "torch", "variance", "variance_epsilon", "weight", "...
deberta/modeling_deberta.py:DebertaSelfOutput
[ -0.0001114058104576543, 0.048472534865140915, 0.03426895663142204, 0.020065374672412872, -0.000665792787913233, 0.05658886581659317, 0.019501740112900734, -0.01961446739733219, 0.0015711301239207387, 0.017021751031279564, 0.013470855541527271, 0.002423626836389303, 0.00391725730150938, -0....
[ "Dropout", "LayerNorm", "Linear", "ModelLayerNorm", "ModelSelfOutput", "Module", "__init__", "class", "config", "def", "dense", "dropout", "forward", "hidden_dropout_prob", "hidden_size", "hidden_states", "input_tensor", "layer_norm_eps", "nn", "return", "self", "super" ]
deberta/modeling_deberta.py:build_relative_position
[ 0, -0.010630759410560131, 0.013105645775794983, -0.030598588287830353, 0.000253113335929811, 0.04882274940609932, 0.03172353655099869, -0.012880655936896801, 0.02283644676208496, 0.04139808937907219, 0.0017928860615938902, 0.015974262729287148, 0.003459215397015214, -0.004218555521219969, ...
[ "Model_relative_position", "None", "arange", "def", "device", "dtype", "k_ids", "key_layer", "key_size", "long", "q_ids", "query_layer", "query_size", "rel_pos_ids", "repeat", "return", "size", "torch", "unsqueeze", "view" ]
deberta/modeling_deberta.py:c2p_dynamic_expand
[ -0.0003349598264321685, -0.001252458430826664, 0.012990615330636501, 0.014738231897354126, -0.001201486331410706, 0.0661764070391655, 0.037748515605926514, -0.03751549869775772, 0.010019667446613312, 0.022485997527837753, -0.010835221968591213, -0.01549553219228983, -0.0006116657750681043, ...
[ "Model_dynamic_expand", "Model_pos", "def", "expand", "query_layer", "relative_pos", "return", "size" ]
deberta/modeling_deberta.py:p2c_dynamic_expand
[ -0.0003611352003645152, 0.0007002723286859691, 0.010089787654578686, 0.013022865168750286, -0.0014005446573719382, 0.053968630731105804, 0.04035915061831474, -0.04200167581439018, 0.013785465620458126, 0.023699268698692322, -0.0009899137075990438, -0.01337483525276184, 0.00006141131598269567...
[ "Model_dynamic_expand", "c2p_pos", "def", "expand", "key_layer", "query_layer", "return", "size" ]
deberta/modeling_deberta.py:pos_dynamic_expand
[ -0.00029201919096522033, -0.005891169887036085, 0.010621517896652222, 0.0019008454401046038, -0.0012478833086788654, 0.05873757600784302, 0.03203867748379707, -0.049218837171792984, 0.013349449262022972, -0.0028295028023421764, -0.0026553794741630554, -0.0035550163593143225, -0.0014582822332...
[ "Model_dynamic_expand", "Model_index", "def", "expand", "key_layer", "p2c_att", "return", "size" ]
deberta/modeling_deberta.py:scaled_size_sqrt
[ -0.0001465174282202497, -0.0010097345802932978, 0.008360319770872593, 0.011015286669135094, -0.00035834984737448394, 0.039316099137067795, 0.06145964935421944, -0.010337422601878643, 0.011806127615272999, 0.04767641797661781, 0.0036152733955532312, -0.01785041205585003, -0.002654966432601213...
[ "Model_size_sqrt", "def", "dtype", "float", "query_layer", "return", "scale_factor", "size", "sqrt", "tensor", "torch" ]
deberta/modeling_deberta.py:build_rpos
[ -0.00021454879606608301, 0.004355881363153458, 0.028500733897089958, 0.003548168344423175, -0.0007139606750570238, 0.0669248029589653, 0.026308370754122734, -0.012519553303718567, 0.012461859732866287, 0.04384728521108627, 0.002163517288863659, 0.00637516425922513, 0.004009718541055918, 0....
[ "Model_relative_position", "Model_rpos", "def", "else", "if", "key_layer", "query_layer", "relative_pos", "return", "size" ]
deberta/modeling_deberta.py:compute_attention_span
[ -0.0001533226459287107, 0.006664660759270191, 0.05898933857679367, -0.014350290410220623, -0.00046439922880381346, 0.05172911286354065, 0.051955994218587875, -0.023595735430717468, 0.01928497664630413, 0.03811618685722351, 0.026545202359557152, 0.015314539894461632, 0.0031763489823788404, ...
[ "Model_attention_span", "def", "key_layer", "max", "max_relative_positions", "min", "query_layer", "return", "size", "tensor", "torch" ]
deberta/modeling_deberta.py:uneven_size_corrected
[ -0.00026711184182204306, 0.019867345690727234, 0.019405314698815346, -0.01848125271499157, -0.0010901051573455334, 0.03234219178557396, 0.04296891391277313, -0.03857961297035217, 0.0188277754932642, 0.026566799730062485, -0.01582457311451435, 0.03534539416432381, -0.000020416911866050214, ...
[ "Model_size_corrected", "def", "dim", "else", "gather", "if", "index", "key_layer", "p2c_att", "pos_dynamic_expand", "pos_index", "query_layer", "relative_pos", "return", "size", "torch", "unsqueeze" ]
deberta/modeling_deberta.py:DisentangledSelfAttention
[ -0.00004979383447789587, 0.049641162157058716, 0.026505235582590103, -0.012803376652300358, -0.0006352552445605397, 0.011287187226116657, 0.022686684504151344, -0.006513998378068209, 0.0014951311750337481, 0.019429685547947884, 0.013589548878371716, 0.025045201182365417, 0.002077740849927067...
[ "Dropout", "False", "Linear", "ModelSelfAttention", "Model_att_bias", "Module", "None", "Parameter", "Relative", "The", "ValueError", "__init__", "a", "all_head_size", "and", "att_span", "attention", "attention_head_size", "attention_mask", "attention_probs", "attention_probs...
deberta/modeling_deberta.py:DebertaEmbeddings
[ -0.0002776140463538468, 0.029344698414206505, 0.017079532146453857, -0.00957141537219286, -0.0013755327090620995, 0.03232501819729805, 0.027854537591338158, -0.005272875539958477, 0.003238233271986246, -0.002923007123172283, 0.0268228892236948, 0.032554276287555695, -0.0017050874885171652, ...
[ "Dropout", "Embedding", "False", "LayerNorm", "Linear", "ModelEmbeddings", "ModelLayerNorm", "Module", "None", "True", "__init__", "arange", "class", "config", "def", "device", "dim", "dropout", "dtype", "else", "embed_proj", "embedding_size", "embeddings", "expand", ...
deberta/modeling_deberta.py:DebertaAttention
[ 0.00003389412086107768, 0.030005263164639473, 0.027989983558654785, -0.0024211336858570576, -0.00019243113638367504, 0.03179662302136421, 0.025414904579520226, -0.018473388627171516, 0.006801566109061241, 0.019257107749581337, 0.007781215477734804, 0.024071386083960533, 0.0012385568115860224...
[ "DisentangledSelfAttention", "False", "ModelAttention", "ModelSelfOutput", "Module", "None", "__init__", "att_matrix", "attention_mask", "attention_output", "class", "config", "def", "else", "forward", "hidden_states", "if", "is", "nn", "output", "output_attentions", "query...
deberta/modeling_deberta.py:DebertaIntermediate
[ -0.00025367451598867774, 0.02240910567343235, 0.04047359153628349, 0.012862369418144226, -0.0009432404185645282, 0.03635763004422188, 0.03452831506729126, -0.018864808604121208, -0.001329111517407, -0.003772961674258113, 0.023209432139992714, -0.02103712037205696, -0.0013719861162826419, 0...
[ "ACT2FN", "Linear", "ModelIntermediate", "Module", "__init__", "class", "config", "def", "dense", "else", "forward", "hidden_act", "hidden_size", "hidden_states", "if", "intermediate_act_fn", "intermediate_size", "isinstance", "nn", "return", "self", "str", "super" ]
deberta/modeling_deberta.py:DebertaOutput
[ -0.00022098780027590692, 0.03763921186327934, 0.04676387086510658, 0.03216441720724106, -0.0012332545593380928, 0.05109808221459389, 0.035814281553030014, -0.02395222708582878, -0.0008875155472196639, 0.017564965412020683, 0.00883951224386692, 0.00306531460955739, 0.001682358793914318, 0.0...
[ "Dropout", "LayerNorm", "Linear", "ModelLayerNorm", "ModelOutput", "Module", "__init__", "class", "config", "def", "dense", "dropout", "forward", "hidden_dropout_prob", "hidden_size", "hidden_states", "input_tensor", "intermediate_size", "layer_norm_eps", "nn", "return", "s...
deberta/modeling_deberta.py:DebertaLayer
[ 0, 0.010092663578689098, 0.027923036366701126, 0.00925160851329565, 0.00018748524598777294, 0.047099098563194275, 0.02668948844075203, 0.007905920036137104, 0.006644336972385645, 0.007625568192452192, 0.0020886207930743694, 0.006307914853096008, 0.0014998819679021835, 0.009419819340109825,...
[ "False", "GradientCheckpointingLayer", "ModelAttention", "ModelIntermediate", "ModelLayer", "ModelOutput", "None", "__init__", "att_matrix", "attention", "attention_mask", "attention_output", "class", "config", "def", "else", "forward", "hidden_states", "if", "intermediate", ...
deberta/modeling_deberta.py:DebertaEncoder
[ -0.00016831814718898386, 0.023245621472597122, 0.034925129264593124, 0.012813439592719078, -0.0006165759405121207, 0.044676948338747025, 0.0053861807100474834, -0.025173306465148926, 0.0026505677960813046, 0.022111687809228897, 0.020297396928071976, 0.01644202508032322, 0.001282761455513537,...
[ "BaseModelOutput", "Embedding", "False", "ModelEncoder", "ModelLayer", "Module", "ModuleList", "None", "True", "_", "__init__", "all_attentions", "all_hidden_states", "and", "att_m", "attention_mask", "attentions", "build_relative_position", "class", "config", "def", "dim",...
deberta/modeling_deberta.py:DebertaPreTrainedModel
[ -0.00029308191733434796, 0.03911571577191353, -0.0046035307459533215, -0.010350795462727547, -0.0013939262134954333, 0.015669161453843117, 0.017270388081669807, -0.005346958059817553, 0, 0.0018585682846605778, -0.0005289771361276507, 0.014125118963420391, -0.0015154479769989848, -0.0040030...
[ "DisentangledSelfAttention", "LegacyModelLMPredictionHead", "Model", "ModelConfig", "ModelEmbeddings", "ModelLMPredictionHead", "ModelPreTrainedModel", "PreTrainedModel", "True", "_init_weights", "_keys_to_ignore_on_load_unexpected", "arange", "base_model_prefix", "bias", "class", "con...
deberta/modeling_deberta.py:DebertaModel
[ -0.0001232814247487113, 0.03561776503920555, 0.01949959807097912, 0.0039731841534376144, -0.0003821724094450474, 0.04779092222452164, 0.010087942704558372, 0, 0, 0.010031585581600666, 0.029080325737595558, 0.0024515390396118164, 0.00025184632977470756, 0.017808882519602776, -0.0435077734...
[ "BaseModelOutput", "False", "ModelEmbeddings", "ModelEncoder", "ModelModel", "ModelPreTrainedModel", "None", "True", "_", "__init__", "and", "append", "attention_mask", "attentions", "auto_docstring", "class", "config", "def", "device", "dtype", "elif", "else", "embedding...
deberta/modeling_deberta.py:LegacyDebertaPredictionHeadTransform
[ -0.0003576379967853427, 0.043465662747621536, 0.04901446774601936, 0.03444884717464447, -0.0014739020261913538, 0.024738434702157974, 0.02601003646850586, -0.011271015740931034, -0.005404307506978512, 0.014912420883774757, 0.012195817194879055, 0.00791861116886139, -0.00024384410062339157, ...
[ "ACT2FN", "LayerNorm", "Linear", "ModelModelPredictionHeadTransform", "Module", "__init__", "class", "config", "def", "dense", "else", "embedding_size", "eps", "forward", "getattr", "hidden_act", "hidden_size", "hidden_states", "if", "isinstance", "layer_norm_eps", "nn", ...
deberta/modeling_deberta.py:LegacyDebertaLMPredictionHead
[ -0.0003152738499920815, 0.03943265229463577, 0.03366764262318611, 0.023405930027365685, -0.0018087710486724973, 0.025135433301329613, 0.05349927023053169, -0.025596633553504944, -0.0015277269994840026, 0.0005008350126445293, 0.017640922218561172, 0.007062134332954884, -0.0001225064042955637,...
[ "Linear", "ModelModelLMPredictionHead", "ModelModelPredictionHeadTransform", "Module", "Parameter", "__init__", "bias", "class", "config", "decoder", "def", "embedding_size", "forward", "getattr", "hidden_size", "hidden_states", "nn", "return", "self", "super", "torch", "tr...
deberta/modeling_deberta.py:LegacyDebertaOnlyMLMHead
[ -0.0003432687954045832, 0.018037870526313782, 0.01584095135331154, 0.03330068662762642, -0.0015609696274623275, 0.03908205404877663, 0.037232015281915665, -0.009423631243407726, -0.0035699952859431505, -0.005983716808259487, 0.019425399601459503, 0.011678365059196949, -0.0022547338157892227,...
[ "ModelModelLMPredictionHead", "ModelModelOnlyMLMHead", "Module", "__init__", "class", "config", "def", "forward", "nn", "prediction_scores", "predictions", "return", "self", "sequence_output", "super" ]
deberta/modeling_deberta.py:DebertaLMPredictionHead
[ -0.0003787095192819834, 0.03589000925421715, 0.04101715236902237, 0.024703513830900192, -0.0018061531009152532, 0.037288323044776917, 0.044046830385923386, -0.009030764922499657, -0.0034520828630775213, -0.007865505293011665, 0.020625103265047073, 0.022023415192961693, 0.002476177643984556, ...
[ "ACT2FN", "LayerNorm", "Linear", "ModelLMPredictionHead", "Module", "Parameter", "True", "__init__", "bias", "class", "config", "def", "dense", "elementwise_affine", "else", "eps", "forward", "hidden_act", "hidden_size", "hidden_states", "if", "isinstance", "layer_norm_ep...
deberta/modeling_deberta.py:DebertaOnlyMLMHead
[ -0.00040845514740794897, 0.018672235310077667, 0.025090817362070084, 0.023573698475956917, -0.0017869913717731833, 0.0385114848613739, 0.04457996413111687, -0.010094677098095417, -0.002436143346130848, -0.011086639948189259, 0.012953863479197025, 0.022873489186167717, -0.001269128522835672, ...
[ "ModelLMPredictionHead", "ModelOnlyMLMHead", "Module", "__init__", "class", "config", "def", "forward", "lm_head", "nn", "prediction_scores", "return", "self", "sequence_output", "super", "word_embeddings" ]