identifier
stringlengths
24
102
embedding
listlengths
2.56k
2.56k
tokens
listlengths
4
448
gpt_oss/modeling_gpt_oss.py:GptOssPreTrainedModel
[ -0.0003385136369615793, 0.04309926927089691, -0.009968145750463009, 0.003543589496985078, -0.0012460182188078761, 0.0447126105427742, 0.03710685670375824, -0.016133416444063187, -0.00939195230603218, 0.0005797946359962225, 0.0030826348811388016, -0.0101986238732934, -0.0029962058179080486, ...
[ "False", "ModelAttention", "ModelConfig", "ModelDecoderLayer", "ModelExperts", "ModelPreTrainedModel", "ModelTopKRouter", "OutputRecorder", "PreTrainedModel", "True", "_can_compile_fullgraph", "_can_record_outputs", "_init_weights", "_keep_in_fp32_modules", "_no_split_modules", "_skip_...
gpt_oss/modeling_gpt_oss.py:GptOssModel
[ -0.00011336923489579931, 0.04792090877890587, -0.0035996928345412016, -0.008999232202768326, -0.0005800286307930946, 0.04184642806649208, 0.03217225521802902, -0.016986049711704254, 0.00680566905066371, -0.0051464359275996685, 0.025647811591625214, -0.002910689217969775, -0.00177172385156154...
[ "DynamicCache", "Embedding", "False", "ModelDecoderLayer", "ModelModel", "ModelPreTrainedModel", "ModelRMSNorm", "ModelRotaryEmbedding", "ModuleList", "MoeModelOutputWithPast", "None", "ValueError", "You", "__init__", "_no_split_modules", "and", "arange", "attention_mask", "atten...
gpt_oss/modeling_gpt_oss.py:load_balancing_loss_func
[ -0.00027690583374351263, 0.01840798556804657, 0.001643570140004158, -0.028812499716877937, -0.0009432663209736347, 0.05808233842253685, 0.038874007761478424, -0.011547867208719254, 0, -0.02252405695617199, 0.031556546688079834, -0.0061455233953893185, -0.0008396499906666577, 0.014349081553...
[ "Model_balancing_loss_func", "None", "_", "attention_mask", "batch_size", "cat", "compute_device", "concatenated_gate_logits", "def", "device", "dim", "else", "expand", "expert_attention_mask", "expert_mask", "float", "for", "functional", "gate_logits", "if", "in", "is", ...
gpt_oss/modeling_gpt_oss.py:GptOssForCausalLM
[ -0.0004233771760482341, 0.03737398609519005, -0.001000046031549573, -0.011270591989159584, -0.0016643102280795574, 0.049286942929029465, 0.03994344547390938, -0.006160867866128683, -0.00519731966778636, 0.005752089899033308, 0.027796901762485504, -0.005810486618429422, 0.00019891427655238658...
[ "GenerationMixin", "Linear", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "MoeCausalLMOutputWithPast", "None", "__init__", "_pp_plan", "_tied_weights_keys", "_tp_plan", "attention_mask", "attentions", "auto_docstring", "aux_loss", "cache_position", "can_return_tuple", ...
gpt_oss/modeling_gpt_oss.py:GptOssForSequenceClassification
[ -0.0002121782599715516, 0.015340753830969334, -0.017272552475333214, 0.0278406273573637, -0.0007244244916364551, 0.027954263612627983, 0.015227118507027626, -0.011079433374106884, 0.0032101948745548725, -0.013863496482372284, 0.02909061498939991, 0.008409005589783192, -0.003437465289607644, ...
[ "GenericForSequenceClassification", "ModelForSequenceClassification", "ModelPreTrainedModel", "class", "pass" ]
gpt_oss/modeling_gpt_oss.py:GptOssForTokenClassification
[ -0.00016310509818140417, 0.024394849315285683, -0.0069496952928602695, -0.024848707020282745, -0.0007446102099493146, 0.035854753106832504, 0.0356278270483017, -0.007885776460170746, 0.0015317696379497647, -0.018381234258413315, 0.040166404098272324, 0.024394849315285683, -0.0042832815088331...
[ "GenericForTokenClassification", "ModelForTokenClassification", "ModelPreTrainedModel", "class", "pass" ]
clipseg/modeling_clipseg.py:contrastive_loss
[ 0.00008122165309032425, 0.037291333079338074, 0.037291333079338074, -0.0017656880663707852, 0.00034784057061187923, 0.061926212161779404, 0.017854638397693634, 0.025877924636006355, 0.023052824661135674, -0.006780242547392845, 0.030059074983000755, -0.031415123492479324, 0.002161202253773808...
[ "Model_loss", "arange", "cross_entropy", "def", "device", "functional", "len", "logits", "nn", "return", "torch" ]
clipseg/modeling_clipseg.py:clipseg_loss
[ 0.00016533616872038692, 0.03712017834186554, 0.07967258244752884, -0.018786432221531868, 0.0008841506205499172, 0.04391045495867729, -0.004498558584600687, -0.012165912427008152, 0.026595251634716988, 0.01935228891670704, 0.033272355794906616, -0.005517099983990192, 0.0037346521858125925, ...
[ "Model_loss", "caption_loss", "contrastive_loss", "def", "image_loss", "return", "similarity", "t" ]
clipseg/modeling_clipseg.py:CLIPSegOutput
[ -0.0000951801921473816, 0.012352273799479008, 0.016018474474549294, 0.008122042752802372, -0.0005076276720501482, 0.043091949075460434, 0.05076276883482933, -0.007783624809235334, 0.020869137719273567, 0.0027778514195233583, 0.020869137719273567, 0.019064240157604218, -0.0022843244951218367,...
[ "Model", "ModelOutput", "None", "class", "def", "else", "for", "getattr", "if", "image_embeds", "in", "k", "keys", "logits_per_image", "logits_per_text", "loss", "not", "r", "return", "self", "text_embeds", "text_model_output", "to_tuple", "tuple", "vision_model_outpu...
clipseg/modeling_clipseg.py:CLIPSegDecoderOutput
[ -0.00012038292334182188, 0.018947381526231766, 0.016892844811081886, 0.019175661727786064, -0.0009416620014235377, 0.05889667570590973, 0.06620169430971146, -0.012498422525823116, 0.021001916378736496, -0.019289802759885788, 0.01609385944902897, 0.004793915431946516, -0.0021401408594101667, ...
[ "Model", "ModelOutput", "None", "attentions", "class", "hidden_states", "logits", "r" ]
clipseg/modeling_clipseg.py:CLIPSegImageSegmentationOutput
[ -0.00014179597201291472, 0.014204257167875767, 0.013584230095148087, 0.01871354505419731, -0.0005037719965912402, 0.04374009370803833, 0.05681702867150307, -0.016909830272197723, 0.01747349090874195, 0.00008366842666873708, 0.016684366390109062, 0.005129314959049225, -0.001831898232921958, ...
[ "Model", "ModelOutput", "None", "class", "conditional_embeddings", "decoder_output", "def", "else", "for", "getattr", "if", "in", "k", "keys", "logits", "loss", "not", "pooled_output", "r", "return", "self", "to_tuple", "tuple", "vision_model_output" ]
clipseg/modeling_clipseg.py:CLIPSegVisionEmbeddings
[ -0.00016093668818939477, 0.03084355965256691, 0.020487328991293907, 0.016434889286756516, -0.0006367112509906292, 0.00810487661510706, 0.023301521316170692, -0.021050166338682175, 0.0009216483449563384, 0.013676980510354042, 0.019811920821666718, 0.018348541110754013, -0.002082503167912364, ...
[ "Conv2d", "Embedding", "False", "Input", "Model", "Module", "Parameter", "True", "ValueError", "_", "__init__", "align_corners", "and", "arange", "batch_size", "bicubic", "cat", "class", "class_embedding", "class_embeds", "class_pos_embed", "config", "def", "dim", "do...
clipseg/modeling_clipseg.py:CLIPSegTextEmbeddings
[ -0.0002546473406255245, 0.017330365255475044, 0.004734288435429335, -0.006685389205813408, -0.0014202866004779935, 0.04039926454424858, 0.0438423827290535, -0.0007244896260090172, 0.00889472384005785, 0.0007854615105316043, 0.020543944090604782, 0.011649219319224358, -0.0009898967109620571, ...
[ "Embedding", "False", "Model", "Module", "None", "Sequence", "ValueError", "__init__", "and", "arange", "be", "class", "config", "def", "else", "embed_dim", "embeddings", "expand", "f", "forward", "got", "hidden_size", "if", "input_ids", "inputs_embeds", "is", "le...
clipseg/modeling_clipseg.py:eager_attention_forward
[ 0.000051929029723396525, 0.031000085175037384, 0.02477744035422802, -0.010239078663289547, 0.00008883178816176951, 0.035978201776742935, 0.05928483232855797, -0.020365018397569656, 0.022175243124365807, 0.01170988567173481, 0.02477744035422802, 0.031452640891075134, 0.0025173425674438477, ...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "softmax", "...
clipseg/modeling_clipseg.py:CLIPSegAttention
[ -0.0001155039863078855, 0.03588911518454552, 0.03837200999259949, 0.006771531421691179, -0.00039324257522821426, 0.017493123188614845, 0.0300204548984766, -0.018847428262233734, 0.0011568032205104828, 0.0388234481215477, 0.012978767976164818, 0, -0.0008746561361476779, -0.01478451024740934...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "Model", "Module", "None", "ValueError", "__init__", "_attn_implementation", "and", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "batch_size", "be", "by", "causal_attention_mask", "cla...
clipseg/modeling_clipseg.py:CLIPSegMLP
[ -0.0001473212760174647, 0.03635108098387718, 0.042485326528549194, 0.011530108749866486, -0.0003603158984333277, 0.02885367162525654, 0.0422581322491169, -0.022037843242287636, 0.006560234352946281, -0.008349388837814331, 0.03021683730185032, -0.019879497587680817, 0.002456537913531065, -0...
[ "ACT2FN", "Linear", "Model", "Module", "__init__", "activation_fn", "class", "config", "def", "fc1", "fc2", "forward", "hidden_act", "hidden_size", "hidden_states", "intermediate_size", "nn", "return", "self", "super" ]
clipseg/modeling_clipseg.py:CLIPSegEncoderLayer
[ -0.00006328244489850476, 0.038475729525089264, 0.024412961676716805, 0.011418965645134449, -0.00009052905807038769, 0.04432583972811699, 0.03240061178803444, -0.002770364983007312, 0.00933767668902874, 0.013444004580378532, 0.01620030589401722, -0.0004183673008810729, 0.001371119637042284, ...
[ "False", "GradientCheckpointingLayer", "LayerNorm", "Model", "ModelAttention", "ModelMLP", "__init__", "attention_mask", "attn_weights", "causal_attention_mask", "class", "config", "def", "embed_dim", "eps", "forward", "hidden_size", "hidden_states", "if", "layer_norm1", "lay...
clipseg/modeling_clipseg.py:CLIPSegPreTrainedModel
[ -0.00011236867430852726, 0.05461117625236511, 0.009888443164527416, 0.0292158555239439, -0.0004810784012079239, 0.0292158555239439, 0.01438319031149149, -0.014720296487212181, 0.00362388975918293, 0.006376922130584717, 0.022248998284339905, 0.01449555903673172, -0.0015169770922511816, -0.0...
[ "LayerNorm", "Linear", "Model", "ModelAttention", "ModelConfig", "ModelMLP", "ModelModel", "ModelTextEmbeddings", "ModelVisionEmbeddings", "None", "PreTrainedModel", "True", "_init_weights", "and", "arange", "base_model_prefix", "bias", "class", "class_embedding", "clip", "co...
clipseg/modeling_clipseg.py:CLIPSegEncoder
[ -0.00006546173244714737, 0.023843886330723763, 0.025755897164344788, 0.032841578125953674, -0.00018715905025601387, 0.03374134749174118, 0.021144578233361244, -0.011472058482468128, 0.007479332387447357, -0.005286144558340311, 0.022269289940595627, -0.013384068384766579, -0.00033917085966095...
[ "BaseModelOutput", "False", "Model", "ModelLayer", "Module", "ModuleList", "None", "_", "__init__", "all_attentions", "attention_mask", "attentions", "can_return_tuple", "causal_attention_mask", "class", "config", "def", "else", "encoder_layer", "encoder_states", "enumerate",...
clipseg/modeling_clipseg.py:CLIPSegTextTransformer
[ -0.000022513155272463337, 0.0483560636639595, 0.028563814237713814, 0.0483560636639595, -0.000288168404949829, 0.021141720935702324, 0.02946346066892147, -0.01686839386820793, 0.0110769122838974, 0.007309637498110533, 0.04610694199800491, 0.008209285326302052, 0.0019679793622344732, -0.002...
[ "BaseModelOutputWithPooling", "LayerNorm", "Model", "ModelEncoder", "ModelTextEmbeddings", "Module", "None", "__init__", "_create_4d_causal_attention_mask", "_prepare_4d_attention_mask", "arange", "argmax", "attention_mask", "attentions", "auto_docstring", "causal_attention_mask", "c...
clipseg/modeling_clipseg.py:CLIPSegTextModel
[ 0.00010727879271144047, 0.052247386425733566, 0.018867112696170807, 0.004688868299126625, 0.00009157945169135928, 0.022551221773028374, 0.025118935853242874, -0.009098636917769909, -0.00021542976901400834, -0.0064751035533845425, 0.016745956614613533, 0.009991754777729511, -0.004409768618643...
[ "Model", "ModelEncoderLayer", "ModelPreTrainedModel", "ModelTextConfig", "ModelTextEmbeddings", "ModelTextTransformer", "None", "__init__", "_no_split_modules", "attention_mask", "auto_docstring", "class", "config", "def", "embeddings", "forward", "get_input_embeddings", "input_ids...
clipseg/modeling_clipseg.py:CLIPSegVisionTransformer
[ -0.00007717513653915375, 0.03567519783973694, 0.009539728984236717, 0.05644809827208519, -0.00018874832312576473, 0.028336945921182632, 0.02585322968661785, -0.02686929516494274, 0.009991313330829144, 0.03386886045336723, 0.03861049935221672, 0.018627872690558434, 0.0016652188496664166, -0...
[ "BaseModelOutputWithPooling", "LayerNorm", "Model", "ModelEncoder", "ModelVisionEmbeddings", "Module", "None", "True", "__init__", "attentions", "auto_docstring", "class", "config", "def", "else", "embed_dim", "embeddings", "encoder", "encoder_outputs", "eps", "forward", "h...
clipseg/modeling_clipseg.py:CLIPSegVisionModel
[ 0, 0.04892557114362717, 0.006841760594397783, 0.03708513453602791, 0.00009293973562307656, 0.006059844978153706, 0.0446808859705925, -0.009997348301112652, 0.003490694100037217, 0.023010656237602234, 0.018989376723766327, 0.010109050199389458, -0.000506150652654469, 0.015191501006484032, ...
[ "Model", "ModelPreTrainedModel", "ModelVisionConfig", "ModelVisionTransformer", "None", "True", "__init__", "auto_docstring", "class", "config", "def", "embeddings", "forward", "get_input_embeddings", "image", "input_modalities", "interpolate_pos_encoding", "kwargs", "main_input_...
clipseg/modeling_clipseg.py:CLIPSegModel
[ -0.00009846658213064075, 0.050191812217235565, 0.03546888008713722, 0.02174978516995907, -0.00042000788380391896, 0.02710357867181301, 0.031230462715029716, -0.008588377386331558, 0.005270140245556831, 0.016284454613924026, 0.04193804785609245, 0.023422846570611, -0.0014290724648162723, 0....
[ "Linear", "Model", "ModelConfig", "ModelOutput", "ModelPreTrainedModel", "ModelTextConfig", "ModelTextTransformer", "ModelVisionConfig", "ModelVisionTransformer", "Model_loss", "None", "Parameter", "True", "TypeError", "__init__", "_attn_implementation", "attention_mask", "auto_doc...
clipseg/modeling_clipseg.py:CLIPSegDecoderLayer
[ -0.00007255468517541885, 0.04213126748800278, 0.026728438213467598, 0.018573999404907227, -0.0001778474688762799, 0.053683388978242874, 0.03148519620299339, -0.015629341825842857, 0.010872584767639637, 0.015969108790159225, 0.02899356000125408, -0.000824645918328315, 0.002420848933979869, ...
[ "False", "LayerNorm", "Model", "ModelAttention", "ModelMLP", "Module", "__init__", "attention_mask", "attn_weights", "causal_attention_mask", "class", "config", "def", "embed_dim", "eps", "forward", "hidden_size", "hidden_states", "if", "layer_norm1", "layer_norm2", "layer_...
clipseg/modeling_clipseg.py:CLIPSegDecoder
[ -0.00029264940530993044, 0.05047845467925072, 0.015760241076350212, 0.033347759395837784, -0.0008957927348092198, 0.029007980599999428, 0.025467636063694954, -0.045910269021987915, 0.004225572105497122, -0.004967902321368456, 0.017587516456842422, -0.0064811138436198235, -0.00144183367956429...
[ "Conv2d", "ConvTranspose2d", "Linear", "Model", "ModelLayer", "ModelOutput", "ModelPreTrainedModel", "ModuleList", "None", "ReLU", "Sequential", "True", "_", "__init__", "activation", "activations", "all_attentions", "all_hidden_states", "attention_mask", "attentions", "batch...
clipseg/modeling_clipseg.py:CLIPSegForImageSegmentation
[ -0.00011369885032763705, 0.05082513764500618, -0.004114149138331413, 0.02451697178184986, -0.0002921185805462301, 0.03671948239207268, 0.02854715846478939, -0.03716728091239929, 0.005037733819335699, -0.007892449386417866, 0.047242749482393265, 0.017352193593978882, -0.0001443100773030892, ...
[ "BCEWithLogitsLoss", "BaseModelOutputWithPooling", "Invalid", "Make", "Model", "ModelConfig", "ModelDecoder", "ModelImageSegmentationOutput", "ModelModel", "ModelPreTrainedModel", "None", "True", "ValueError", "__init__", "activations", "are", "as", "attention_mask", "attentions"...
gemma/modeling_gemma.py:GemmaRMSNorm
[ -0.00007788453513057902, 0.04100974649190903, 0.021411165595054626, 0.05097896605730057, -0.00047438760520890355, 0.053697843104600906, 0.0159734096378088, -0.010195793583989143, 0.009855933487415314, 0.04191603884100914, 0.013990893959999084, 0.012518168427050114, 0.0031578638590872288, 0...
[ "ModelRMSNorm", "Module", "Parameter", "True", "__init__", "_norm", "class", "def", "dim", "eps", "extra_repr", "f", "float", "forward", "keepdim", "mean", "nn", "output", "pow", "return", "rsqrt", "self", "shape", "super", "torch", "tuple", "type_as", "weight",...
gemma/modeling_gemma.py:GemmaMLP
[ -0.0002434849739074707, 0.024471141397953033, 0.0228551235049963, 0.026548879221081734, -0.0009450823999941349, 0.060485273599624634, 0.036244992166757584, -0.005078915972262621, 0, -0.005021201446652412, 0.028164898976683617, -0.04640282690525055, -0.0015583038330078125, 0.009176678024232...
[ "ACT2FN", "Linear", "ModelMLP", "Module", "__init__", "act_fn", "class", "config", "def", "down_proj", "forward", "gate_proj", "hidden_act", "hidden_size", "intermediate_size", "nn", "return", "self", "super", "up_proj", "x" ]
gemma/modeling_gemma.py:GemmaRotaryEmbedding
[ -0.00029968636226840317, 0.04852752387523651, 0.0020364229567348957, -0.005228263325989246, -0.0013792794197797775, 0.04136393964290619, 0.040670689195394516, 0.0062681385315954685, -0.0026430170983076096, 0.021144136786460876, 0.0021808501332998276, -0.004043960478156805, -0.001364836702123...
[ "False", "ModelRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "arange", "attention_factor", "attention_scaling", "base", "cat", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cpu", "def", "default", "device", ...
gemma/modeling_gemma.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
gemma/modeling_gemma.py:apply_rotary_pos_emb
[ -0.0001444592053303495, 0.027112245559692383, 0.028019767254590988, 0.0028360087890177965, -0.0005707467789761722, 0.021667107939720154, 0.046510547399520874, -0.002183726755902171, 0.01293220091611147, 0.03652779385447502, 0.007884104736149311, 0.0017654155381023884, -0.0007834474672563374,...
[ "Model_rotary_pos_emb", "cos", "def", "k", "k_embed", "q", "q_embed", "return", "rotate_half", "sin", "unsqueeze", "unsqueeze_dim" ]
gemma/modeling_gemma.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
gemma/modeling_gemma.py:eager_attention_forward
[ 0, 0.021594731137156487, 0.01775064319372177, -0.018768195062875748, -0.00008832923776935786, 0.039797618985176086, 0.05359111353754997, -0.035049039870500565, 0.02069023996591568, 0.011645326390862465, 0.029395969584584236, 0.022499222308397293, 0.002741739386692643, -0.014584923163056374...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "causal_mask", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "key_states", "kwargs", "matmul", "module", "nn", "not", "num_key_value_groups", ...
gemma/modeling_gemma.py:GemmaAttention
[ -0.00014813605230301619, 0.03160235658288002, 0.03182808682322502, -0.010835093446075916, -0.000684247468598187, 0.03227955102920532, 0.04018013924360275, -0.008239186368882656, 0.0032166684977710247, 0.00919854361563921, 0.01839708723127842, 0.025733347982168198, -0.0009311408502981067, -...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "ModelAttention", "Module", "None", "Tensor", "__init__", "_attn_implementation", "apply_rotary_pos_emb", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "cache_kwargs", "cache_position", "cl...
gemma/modeling_gemma.py:GemmaDecoderLayer
[ -0.00015529245138168335, 0.04223954677581787, 0.012649276293814182, 0.003317611524835229, -0.0006635222816839814, 0.03862546756863594, 0.04246542602777481, -0.03501138836145401, 0.00671992776915431, -0.0017364519881084561, -0.003839958691969514, 0.021684473380446434, -0.0015670419670641422, ...
[ "False", "GradientCheckpointingLayer", "ModelAttention", "ModelDecoderLayer", "ModelMLP", "ModelRMSNorm", "None", "Tensor", "_", "__init__", "attention_mask", "cache_position", "class", "config", "def", "eps", "forward", "hidden_size", "hidden_states", "input_layernorm", "kwa...
gemma/modeling_gemma.py:GemmaPreTrainedModel
[ -0.00024435759405605495, 0.03762398660182953, -0.001232412178069353, 0.013372380286455154, -0.0009420161950401962, 0.026518110185861588, 0.019265294075012207, -0.022778376936912537, -0.005099636502563953, 0.011332525871694088, 0.00010535707406233996, 0.006431208457797766, -0.0029322910122573...
[ "ModelAttention", "ModelConfig", "ModelDecoderLayer", "ModelPreTrainedModel", "PreTrainedModel", "RMSNorm", "True", "__class__", "__name__", "_can_compile_fullgraph", "_can_record_outputs", "_init_weights", "_no_split_modules", "_skip_keys_device_placement", "_supports_attention_backend"...
gemma/modeling_gemma.py:GemmaModel
[ -0.00007910186832305044, 0.05062519758939743, -0.005990648176521063, 0.001954694977030158, -0.0005765647510997951, 0.03622514009475708, 0.04185016080737114, -0.01473755668848753, 0.01243129838258028, 0.009956288151443005, 0.01800007000565529, 0.00804378092288971, -0.0015257871709764004, 0....
[ "BaseModelOutputWithPast", "DynamicCache", "Embedding", "False", "ModelDecoderLayer", "ModelModel", "ModelPreTrainedModel", "ModelRMSNorm", "ModelRotaryEmbedding", "ModuleList", "None", "ValueError", "You", "__init__", "and", "arange", "attention_mask", "auto_docstring", "cache_p...
gemma/modeling_gemma.py:GemmaForCausalLM
[ -0.00028262686100788414, 0.03526614233851433, 0.008361488580703735, -0.001301150070503354, -0.0012087186332792044, 0.02741658128798008, 0.0389065183699131, -0.007394513580948114, 0.003640376031398773, 0.026734011247754097, 0.025823917239904404, 0.0026734010316431522, 0.0009385344455949962, ...
[ "CausalLMOutputWithPast", "GenerationMixin", "Linear", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "_pp_plan", "_tied_weights_keys", "_tp_plan", "attention_mask", "attentions", "auto_docstring", "cache_position", "can_return_tuple", "class", "colwi...
gemma/modeling_gemma.py:GemmaForSequenceClassification
[ -0.0002121782599715516, 0.015340753830969334, -0.017272552475333214, 0.0278406273573637, -0.0007244244916364551, 0.027954263612627983, 0.015227118507027626, -0.011079433374106884, 0.0032101948745548725, -0.013863496482372284, 0.02909061498939991, 0.008409005589783192, -0.003437465289607644, ...
[ "GenericForSequenceClassification", "ModelForSequenceClassification", "ModelPreTrainedModel", "class", "pass" ]
gemma/modeling_gemma.py:GemmaForTokenClassification
[ -0.00016310509818140417, 0.024394849315285683, -0.0069496952928602695, -0.024848707020282745, -0.0007446102099493146, 0.035854753106832504, 0.0356278270483017, -0.007885776460170746, 0.0015317696379497647, -0.018381234258413315, 0.040166404098272324, 0.024394849315285683, -0.0042832815088331...
[ "GenericForTokenClassification", "ModelForTokenClassification", "ModelPreTrainedModel", "class", "pass" ]
patchtst/modeling_patchtst.py:eager_attention_forward
[ 0.00002095530362566933, 0.025862814858555794, 0.0269921962171793, -0.01146321278065443, 0.00008955634984886274, 0.03704368323087692, 0.060083046555519104, -0.023942869156599045, 0.021458230912685394, 0.014569009654223919, 0.023152301087975502, 0.026879258453845978, 0.003063444746658206, -0...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "functional", "if", "is", "key", "kwargs", "matmul", "module", "nn", "not", "p", "query", "return", "scaling", "shape", "size", "softmax", "tor...
patchtst/modeling_patchtst.py:PatchTSTAttention
[ -0.00012168648390797898, 0.04424449801445007, 0.04559892416000366, -0.005925602745264769, -0.0006137231248430908, 0.02426675334572792, 0.03656943142414093, -0.019526271149516106, -0.001551943481899798, 0.028894366696476936, 0.01636595092713833, 0.0032872986048460007, -0.0034989272244274616, ...
[ "ALL_ATTENTION_FUNCTIONS", "False", "Linear", "ModelAttention", "Module", "None", "ValueError", "__init__", "_attn_implementation", "and", "attention_interface", "attention_mask", "attn_output", "attn_weights", "be", "bsz", "by", "class", "config", "contiguous", "current_stat...
patchtst/modeling_patchtst.py:PatchTSTBatchNorm
[ -0.00005249693276709877, 0.011583590880036354, 0.04318542778491974, 0.011977207846939564, 0.0000896181954885833, 0.029015207663178444, 0.00024073907115962356, -0.017206693068146706, 0.012258362956345081, 0.030139828100800514, 0.026203656569123268, -0.0014620065921917558, -0.00170098850503563...
[ "BatchNorm1d", "ModelBatchNorm", "Module", "__init__", "batchnorm", "class", "config", "d_model", "def", "eps", "forward", "inputs", "nn", "norm_eps", "output", "return", "self", "super", "transpose" ]
patchtst/modeling_patchtst.py:random_masking
[ 0.00004487325713853352, 0.016026537865400314, 0.024432066828012466, -0.017707644030451775, 0.000383502192562446, 0.03025989793241024, -0.00941419042646885, -0.05289878323674202, 0.012216033414006233, 0.0052674636244773865, 0.014569580554962158, 0.002675759606063366, -0.0014219350414350629, ...
[ "False", "Mask", "Model_masking", "None", "ValueError", "and", "argsort", "batch_size", "be", "between", "bool", "channel_consistent_masking", "def", "device", "dim", "else", "f", "gather", "has", "ids_restore", "ids_shuffle", "if", "index", "inputs", "inputs_mask", ...
patchtst/modeling_patchtst.py:forecast_masking
[ 0.00015528762014582753, 0.012987691909074783, 0.007340869400650263, -0.013947651721537113, 0.0007552625029347837, 0.03342919051647186, -0.00703029427677393, -0.042238231748342514, 0.01264888234436512, -0.007623210549354553, 0.01208420004695654, -0.003077518194913864, -0.0023998995311558247, ...
[ "Model_mask_ratios", "Model_masking", "None", "ValueError", "_", "and", "append", "batch1", "batch2", "batch_size", "be", "bool", "def", "device", "elif", "f", "for", "greater", "if", "in", "inputs", "inputs_mask", "int", "is", "isinstance", "key", "lambda", "le...
patchtst/modeling_patchtst.py:PatchTSTPatchify
[ -0.0002431098255328834, 0.011470524594187737, -0.005280984099954367, 0.011640879325568676, -0.0006246325210668147, 0.040430761873722076, 0.007126489654183388, -0.027029553428292274, -0.0034070867113769054, 0.008120222948491573, 0.030890919268131256, -0.02930094487965107, 0.000068319182901177...
[ "Input", "Model", "ModelModelify", "Model_length", "Model_stride", "Module", "Sequence", "ValueError", "__init__", "be", "class", "config", "configuration", "context_length", "contiguous", "def", "dimension", "doesn", "f", "forward", "greater", "has", "if", "length", ...
patchtst/modeling_patchtst.py:PatchTSTMasking
[ -0.00002322904583706986, 0.025245150551199913, 0.02176693081855774, 0.00048035912914201617, 0, 0.04286065697669983, -0.005077080335468054, -0.04039224237203598, 0.009368755854666233, -0.001157069462351501, 0.02221573330461979, 0.004375826101750135, 0.000575028476305306, -0.0210937261581420...
[ "ModelMasking", "Model_input", "Module", "None", "__init__", "bool", "channel_consistent_masking", "class", "config", "def", "elif", "else", "forecast", "forecast_masking", "forward", "if", "inputs", "is", "mask", "mask_ratio", "mask_type", "mask_value", "masked_input", ...
patchtst/modeling_patchtst.py:PatchTSTEncoderLayer
[ -0.0001732015225570649, 0.03592643886804581, 0.019213823601603508, 0.014381944201886654, -0.000497399247251451, 0.02842281386256218, 0.004661341663450003, -0.029559727758169174, 0.0009308471926487982, 0.022169794887304306, 0.027058519423007965, 0.012392346747219563, 0.0010232212953269482, ...
[ "ACT2CLS", "Dropout", "Identity", "LayerNorm", "Linear", "ModelAttention", "ModelBatchNorm", "ModelEncoderLayer", "Module", "None", "Sequential", "_", "__init__", "activation_function", "attention_dropout", "attn_output", "attn_weights", "batch_size", "batchnorm", "channel_atte...
patchtst/modeling_patchtst.py:PatchTSTPreTrainedModel
[ -0.00032766000367701054, 0.034606628119945526, -0.008880839683115482, 0.0038388145621865988, -0.001246182364411652, 0.027158180251717567, 0.0069900802336633205, -0.011115373112261295, -0.006818193010985851, 0.007734925020486116, 0.001876435475423932, 0.0016830622917041183, -0.001167400740087...
[ "BatchNorm1d", "False", "GatheredParameters", "LayerNorm", "Linear", "ModelConfig", "ModelEncoder", "ModelPositionalEncoding", "ModelPreTrainedModel", "Model_length", "Model_stride", "None", "PreTrainedModel", "True", "_init_pe", "_init_weights", "_set_gradient_checkpointing", "_su...
patchtst/modeling_patchtst.py:PatchTSTEmbedding
[ -0.00013976116315461695, 0.016643963754177094, 0.017776206135749817, 0.04076072573661804, -0.0002989827480632812, 0.03781689703464508, -0.007104821037501097, -0.01834232732653618, -0.0007041132193990052, -0.008605042472481728, 0.031249890103936195, -0.015285272151231766, -0.00202388316392898...
[ "Linear", "ModelEmbedding", "Model_input", "Model_length", "Module", "ModuleList", "The", "ValueError", "_", "__init__", "append", "as", "batch", "be", "channels", "class", "config", "d_model", "def", "defined", "dim", "else", "embeddings", "f", "for", "forward", ...
patchtst/modeling_patchtst.py:PatchTSTPositionalEncoding
[ -0.00016108076670207083, 0.048940230160951614, 0.006117528770118952, 0.02707572840154171, -0.0010762319434434175, 0.014217589981853962, 0.015407109633088112, 0.0037384897004812956, -0.0004602307453751564, 0.010932250879704952, 0.012461632490158081, 0.00860985554754734, 0.0006620242493227124,...
[ "Available", "Dropout", "False", "Identity", "ModelPositionalEncoding", "Model_input", "Module", "Parameter", "True", "ValueError", "__init__", "_init_pe", "a", "and", "arange", "are", "cat", "class", "cls_token", "cls_tokens", "config", "cos", "d_model", "def", "dim"...
patchtst/modeling_patchtst.py:PatchTSTEncoder
[ -0.00005459879321279004, 0.03697218745946884, 0.01814793422818184, 0.02547474019229412, -0.00037338529364205897, 0.04305907338857651, 0.0055232844315469265, -0.008566726930439472, 0.005466924048960209, 0.00464970339089632, -0.00029765148065052927, 0.0062841447070240974, -0.000676320516504347...
[ "BaseModelOutput", "False", "ModelEmbedding", "ModelEncoder", "ModelEncoderLayer", "ModelPositionalEncoding", "ModelPreTrainedModel", "Model_input", "ModuleList", "None", "__init__", "all_attentions", "attentions", "class", "config", "def", "else", "embedder", "encoder_layer", ...
patchtst/modeling_patchtst.py:PatchTSTModelOutput
[ -0.00020307977683842182, 0.018402982503175735, 0.00408316170796752, 0.015872571617364883, -0.0011573750525712967, 0.062110062688589096, 0.05405876040458679, -0.019208112731575966, 0.01633264683187008, -0.004830782767385244, 0.011386845260858536, 0.02265867218375206, -0.0021422221325337887, ...
[ "ModelModelOutput", "ModelOutput", "Model_input", "None", "attentions", "class", "hidden_states", "last_hidden_state", "loc", "mask", "r", "scale" ]
patchtst/modeling_patchtst.py:PatchTSTForPretrainingOutput
[ -0.0001838255557231605, 0.02963833697140217, 0.04027193784713745, 0.02138032577931881, -0.0010463916696608067, 0.04615435749292374, 0.05656171217560768, -0.005316800903528929, 0.015611032024025917, -0.02963833697140217, 0.0009261980303563178, 0.025792140513658524, -0.0007317671552300453, 0...
[ "ModelForPretrainingOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "loss", "prediction_output", "r" ]
patchtst/modeling_patchtst.py:PatchTSTForRegressionOutput
[ -0.00033972697565332055, 0.023592954501509666, 0.036083340644836426, 0.02532772906124592, -0.0015396137023344636, 0.06013890355825424, 0.038165073841810226, 0.0041634622029960155, 0.008326924405992031, -0.014919073320925236, 0.01526602916419506, 0.019776446744799614, -0.001257712603546679, ...
[ "ModelForRegressionOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "loss", "r", "regression_outputs" ]
patchtst/modeling_patchtst.py:PatchTSTForPredictionOutput
[ -0.00034172023879364133, 0.00205032154917717, 0.0034026612993329763, 0.022451747208833694, -0.0016649774042889476, 0.04676477983593941, 0.07352074980735779, 0, 0.010818718001246452, -0.009306423366069794, 0.003751652082428336, 0.017333215102553368, -0.001737683778628707, 0.0048858723603188...
[ "ModelForPredictionOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "loc", "loss", "prediction_outputs", "r", "scale" ]
patchtst/modeling_patchtst.py:PatchTSTForClassificationOutput
[ -0.0002548995253164321, 0.01426011323928833, 0.026694931089878082, 0.02178945206105709, -0.0011764593655243516, 0.02361474744975567, 0.05475883185863495, -0.0006951804971322417, 0.015172760002315044, -0.03422427177429199, 0.02441331371665001, 0.015743164345622063, -0.0009482975001446903, -...
[ "ModelForClassificationOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "loss", "prediction_logits", "r" ]
patchtst/modeling_patchtst.py:SamplePatchTSTOutput
[ -0.000104561957414262, 0.010593239217996597, -0.004214514512568712, 0.028704259544610977, -0.00044672429794445634, 0.05308010056614876, 0.050346363335847855, -0.02403412200510502, 0.01685805805027485, -0.0011746535310521722, 0.0252870861440897, 0.00020734414283651859, -0.0010465094819664955,...
[ "ModelModelOutput", "ModelOutput", "None", "class", "r", "sequences" ]
patchtst/modeling_patchtst.py:nll
[ 0.00003988773823948577, 0.028024237602949142, 0.02416665107011795, 0.009019947610795498, -0.00017018768994603306, 0.06262906640768051, 0.04243346303701401, 0.025414694100618362, 0.01667839288711548, -0.020082145929336548, 0.025301234796643257, -0.028931906446814537, 0.001205496140755713, 0...
[ "Model", "def", "input", "log_prob", "return", "target" ]
patchtst/modeling_patchtst.py:weighted_average
[ -0.00007938854105304927, 0.033500220626592636, 0.03528689965605736, 0.005471702665090561, -0.0003873462846968323, 0.03484022989869118, 0.004159610718488693, -0.008933392353355885, 0.004103776998817921, -0.0012492790119722486, 0.01133424136787653, -0.015521768480539322, -0.0021775143686681986...
[ "Model_average", "Model_tensor", "None", "clamp", "def", "dim", "else", "if", "input_tensor", "is", "min", "not", "return", "sum", "sum_weights", "torch", "weights", "where", "zeros_like" ]
patchtst/modeling_patchtst.py:PatchTSTStdScaler
[ -0.0002436667709844187, 0.061923377215862274, -0.014057972468435764, 0.022993605583906174, -0.0008928519673645496, 0.037108492106199265, 0.008366485126316547, -0.0015224727103486657, 0.008594145067036152, 0.02766062505543232, 0.028457432985305786, 0.02811594493687153, 0.0031445464119315147, ...
[ "ModelStdScaler", "Module", "True", "__init__", "clamp_min", "class", "config", "data", "def", "denominator", "dim", "else", "forward", "hasattr", "if", "keepdim", "loc", "minimum_scale", "nn", "observed_indicator", "return", "scale", "scaling_dim", "self", "sqrt", ...
patchtst/modeling_patchtst.py:PatchTSTMeanScaler
[ -0.0002414414193481207, 0.05749146267771721, -0.0013563326792791486, 0.029995545744895935, -0.001086486387066543, 0.03158621862530708, 0.002783677540719509, 0.005822998937219381, 0.002556438557803631, 0.030222784727811813, 0.01590672880411148, 0.007072813343256712, 0.002954106777906418, 0....
[ "ModelMeanScaler", "Module", "None", "True", "__init__", "abs", "batch_observations", "batch_sum", "clamp", "class", "config", "data", "def", "default_scale", "dim", "else", "forward", "hasattr", "if", "is", "keepdim", "min", "minimum_scale", "nn", "not", "num_obser...
patchtst/modeling_patchtst.py:PatchTSTNOPScaler
[ -0.00023304640490096062, 0.0414431206882, -0.010189997963607311, 0.04121541231870651, -0.0010531562147662044, 0.033473290503025055, 0.017192063853144646, -0.008823741227388382, 0.010588490404188633, 0.03324558213353157, 0.003586424048990011, -0.0070874569937586784, 0.0005479258834384382, 0...
[ "False", "ModelNOPScaler", "Module", "None", "True", "__init__", "class", "config", "data", "def", "dim", "else", "forward", "hasattr", "if", "keepdim", "loc", "mean", "nn", "observed_indicator", "ones_like", "requires_grad", "return", "scale", "scaling_dim", "self"...
patchtst/modeling_patchtst.py:PatchTSTScaler
[ -0.0002990240172948688, 0.049734678119421005, -0.02440895512700081, 0.03965022787451744, -0.0010170397581532598, 0.04675517976284027, 0.017418596893548965, -0.00312274182215333, 0.0026500332169234753, 0.023606782779097557, 0.020627286285161972, 0.010199046693742275, 0.002850576303899288, -...
[ "ModelMeanScaler", "ModelNOPScaler", "ModelScaler", "ModelStdScaler", "Module", "True", "__init__", "class", "config", "data", "def", "elif", "else", "forward", "if", "is", "loc", "mean", "nn", "observed_indicator", "or", "return", "scale", "scaler", "scaling", "sel...
patchtst/modeling_patchtst.py:PatchTSTModel
[ -0.00010400922474218532, 0.02933412790298462, -0.010097709484398365, 0.019856948405504227, -0.0005923237185925245, 0.05054495856165886, 0.028882833197712898, -0.01963130012154579, 0.0065155611373484135, 0.0011493901256471872, 0.017938947305083275, -0.0004583457484841347, 0.001234007766470313...
[ "Identity", "ModelEncoder", "ModelMasking", "ModelModel", "ModelModelOutput", "ModelModelify", "ModelPreTrainedModel", "ModelScaler", "Model_input", "Modeled_values", "Modelifier", "None", "__init__", "attentions", "class", "config", "def", "do_mask_input", "else", "encoder", ...
patchtst/modeling_patchtst.py:PatchTSTMaskPretrainHead
[ -0.00031448935624212027, 0.01771203987300396, 0.040944717824459076, 0.008625993505120277, -0.0014376656617969275, 0.042784929275512695, 0.02081739902496338, -0.03013347089290619, -0.005405622534453869, 0.012881483882665634, 0.017482014372944832, 0.01989729143679142, -0.0006002254085615277, ...
[ "Dropout", "Identity", "Linear", "ModelMaskPretrainHead", "Model_length", "Module", "__init__", "class", "config", "d_model", "def", "dropout", "else", "embedding", "forward", "head_dropout", "if", "linear", "nn", "return", "self", "super", "use_cls_token" ]
patchtst/modeling_patchtst.py:PatchTSTForPretraining
[ -0.00027391233015805483, 0.028344590216875076, 0.026637084782123566, 0.011952538043260574, -0.0012165976222604513, 0.048948489129543304, 0.01946556195616722, -0.029824428260326385, 0.0022909031249582767, -0.004439514130353928, 0.015139881521463394, 0.011098785325884819, 0.0001129443699028343...
[ "MSELoss", "ModelForPretraining", "ModelForPretrainingOutput", "ModelMaskPretrainHead", "ModelModel", "ModelPreTrainedModel", "Model_input", "None", "True", "__init__", "attentions", "class", "config", "def", "dim", "do_mask_input", "else", "encoder_states", "forward", "head", ...
patchtst/modeling_patchtst.py:PatchTSTClassificationHead
[ -0.00034063143539242446, 0.019505631178617477, 0.02868475206196308, 0.03763439506292343, -0.0011617324780672789, 0.02283306233584881, 0.024095192551612854, -0.008892273530364037, -0.0031983498483896255, 0.005249309819191694, 0.02868475206196308, -0.0006740916869603097, 0, -0.00032270347583...
[ "Dropout", "Flatten", "Identity", "Linear", "ModelClassificationHead", "Module", "__init__", "class", "config", "d_model", "def", "dim", "dropout", "elif", "else", "embedding", "flatten", "forward", "head_dropout", "if", "linear", "max", "mean", "nn", "num_input_chann...
patchtst/modeling_patchtst.py:PatchTSTForClassification
[ -0.00032971386099234223, 0.023968765512108803, 0.006307569798082113, 0.013073871843516827, -0.0014120355481281877, 0.017775878310203552, 0.021560419350862503, -0.012729822658002377, -0.00022936616733204573, -0.0012901846785098314, 0.030046967789530754, 0.004529981873929501, 0.001318855443969...
[ "CrossEntropyLoss", "False", "ModelClassificationHead", "ModelForClassification", "ModelForClassificationOutput", "ModelModel", "ModelPreTrainedModel", "None", "True", "__init__", "attentions", "auto_docstring", "class", "config", "def", "do_mask_input", "else", "forward", "head"...
patchtst/modeling_patchtst.py:PatchTSTPredictionHead
[ -0.00047846566303633153, 0.015782006084918976, 0.02543964982032776, 0.037688370794057846, -0.0016562272794544697, 0.02744184620678425, 0.03509729728102684, -0.02473299391567707, -0.009834309108555317, 0.004769934341311455, 0.013720923103392124, -0.015193124301731586, -0.0047993785701692104, ...
[ "Dropout", "Flatten", "Identity", "Linear", "ModelPredictionHead", "Module", "ModuleList", "None", "__init__", "append", "class", "config", "d_model", "def", "dim", "distribution_output", "dropout", "dropouts", "elif", "else", "embedding", "flatten", "flattens", "for", ...
patchtst/modeling_patchtst.py:PatchTSTForPrediction
[ -0.00045970489736646414, 0.022713100537657738, 0.0020741885527968407, 0.026479002088308334, -0.0019712145440280437, 0.025184473022818565, 0.04189566522836685, -0.011062338948249817, -0.011709603480994701, 0.006207854952663183, 0.01588740199804306, -0.006708014290779829, -0.001677003572694957...
[ "False", "MSELoss", "ModelForPrediction", "ModelForPredictionOutput", "ModelModel", "ModelPreTrainedModel", "ModelPredictionHead", "Modelifier", "NegativeBinomialOutput", "None", "NormalOutput", "SampleModelOutput", "StudentTOutput", "True", "_", "__init__", "attentions", "class", ...
patchtst/modeling_patchtst.py:PatchTSTRegressionHead
[ -0.00045175370178185403, 0.02115529589354992, 0.03290823847055435, 0.03925482556223869, -0.0017776325112208724, 0.038784708827733994, 0.024446120485663414, -0.0023358971811830997, -0.014279824681580067, 0.014632413163781166, 0.028089532628655434, -0.013222060166299343, -0.001579301548190415,...
[ "Dropout", "Flatten", "Identity", "Linear", "ModelRegressionHead", "Module", "None", "__init__", "class", "config", "d_model", "def", "dim", "distribution_output", "dropout", "elif", "else", "embedding", "flatten", "forward", "get_parameter_projection", "head_dim", "head_...
patchtst/modeling_patchtst.py:PatchTSTForRegression
[ -0.00046354197547771037, 0.03702402487397194, 0.012816008180379868, 0.02052934654057026, -0.0022843347396701574, 0.052925366908311844, 0.010086673311889172, 0.000778750516474247, -0.012697341851890087, 0.015189343132078648, 0.02349601499736309, 0.005191669799387455, -0.0001335000852122903, ...
[ "False", "MSELoss", "ModelForRegression", "ModelForRegressionOutput", "ModelModel", "ModelPreTrainedModel", "ModelRegressionHead", "NegativeBinomialOutput", "None", "NormalOutput", "SampleModelOutput", "StudentTOutput", "True", "_", "__init__", "attentions", "auto_docstring", "clas...
tapas/modeling_tapas.py:TableQuestionAnsweringOutput
[ -0.0002222355833509937, 0.031176984310150146, 0.011094000190496445, 0.018319323658943176, -0.001031173043884337, 0.05416147783398628, 0.05620960146188736, 0.013938616029918194, 0.014223077334463596, 0.004010907839983702, 0.02332584746181965, 0.020595015957951546, -0.0015076461713761091, -0...
[ "ModelOutput", "ModelQuestionAnsweringOutput", "None", "attentions", "class", "hidden_states", "logits", "logits_aggregation", "loss", "r" ]
tapas/modeling_tapas.py:TapasEmbeddings
[ -0.00026421822258271277, 0.03016372211277485, 0.011254267767071724, -0.006141288205981255, -0.0012782448902726173, 0.03953275829553604, 0.021023200824856758, -0.005512877367436886, 0.006255544722080231, 0.005598569754511118, 0.028564130887389183, 0.026507513597607613, -0.000999744632281363, ...
[ "Dropout", "Embedding", "IndexMap", "LayerNorm", "ModelEmbeddings", "Module", "None", "ProductIndexMap", "__init__", "arange", "as_tensor", "batch_dims", "class", "col_index", "config", "def", "device", "dropout", "dtype", "else", "embeddings", "enumerate", "eps", "expa...
tapas/modeling_tapas.py:TapasSelfAttention
[ -0.000023558890461572446, 0.038420189172029495, 0.031272247433662415, -0.0036018926184624434, -0.00040311881457455456, 0.010107636451721191, 0.018204914405941963, -0.02468273788690567, 0.003071381477639079, 0.016529615968465805, 0.005249269772320986, 0.014407570473849773, 0.00110290513839572...
[ "Dropout", "EncoderDecoderCache", "False", "Linear", "ModelSelfAttention", "Module", "None", "The", "True", "ValueError", "_", "__init__", "a", "all_head_size", "and", "attention", "attention_head_size", "attention_mask", "attention_probs", "attention_probs_dropout_prob", "at...
tapas/modeling_tapas.py:TapasSelfOutput
[ -0.00012168083776487038, 0.04875698313117027, 0.03859927877783775, 0.020879726856946945, -0.0006630723946727812, 0.05620596557855606, 0.02370131015777588, -0.01952536590397358, 0.0035551965702325106, 0.017155233770608902, 0.01365646906197071, 0.003512872848659754, 0.0036962758749723434, -0...
[ "Dropout", "LayerNorm", "Linear", "ModelSelfOutput", "Module", "__init__", "class", "config", "def", "dense", "dropout", "eps", "forward", "hidden_dropout_prob", "hidden_size", "hidden_states", "input_tensor", "layer_norm_eps", "nn", "return", "self", "super" ]
tapas/modeling_tapas.py:TapasAttention
[ 0.00012378250539768487, 0.03438520058989525, 0.03348623961210251, 0, 0.000530245597474277, 0.04090265557169914, 0.04045317694544792, -0.026968784630298615, 0.012416877783834934, -0.0010253756772726774, 0.016293639317154884, 0.03236253932118416, 0.0025423697661608458, -0.009832369163632393,...
[ "False", "ModelAttention", "ModelSelfAttention", "ModelSelfOutput", "Module", "None", "__init__", "attention_mask", "attention_output", "cache_position", "class", "config", "def", "encoder_hidden_states", "forward", "hidden_states", "layer_idx", "nn", "output", "output_attentio...
tapas/modeling_tapas.py:TapasIntermediate
[ -0.00025367451598867774, 0.02240910567343235, 0.04047359153628349, 0.012862369418144226, -0.0009432404185645282, 0.03635763004422188, 0.03452831506729126, -0.018864808604121208, -0.001329111517407, -0.003772961674258113, 0.023209432139992714, -0.02103712037205696, -0.0013719861162826419, 0...
[ "ACT2FN", "Linear", "ModelIntermediate", "Module", "__init__", "class", "config", "def", "dense", "else", "forward", "hidden_act", "hidden_size", "hidden_states", "if", "intermediate_act_fn", "intermediate_size", "isinstance", "nn", "return", "self", "str", "super" ]
tapas/modeling_tapas.py:TapasOutput
[ -0.0002591986849438399, 0.03935529664158821, 0.0507957898080349, 0.030660521239042282, -0.0012513039400801063, 0.04988054931163788, 0.03798243775963783, -0.023681821301579475, 0.002259497530758381, 0.014758236706256866, 0.011040075682103634, 0.00267421524040401, 0.0013013561256229877, 0.00...
[ "Dropout", "LayerNorm", "Linear", "ModelOutput", "Module", "__init__", "class", "config", "def", "dense", "dropout", "eps", "forward", "hidden_dropout_prob", "hidden_size", "hidden_states", "input_tensor", "intermediate_size", "layer_norm_eps", "nn", "return", "self", "su...
tapas/modeling_tapas.py:TapasLayer
[ -0.000011653847650450189, 0.016919855028390884, 0.018376531079411507, 0.00616286089643836, 0.0002976381510961801, 0.03921820595860481, 0.021177830174565315, -0.0038097684737294912, 0.0030534174293279648, 0.001974916784092784, 0.0023250794038176537, 0.018040373921394348, 0.002199020702391863,...
[ "False", "GradientCheckpointingLayer", "If", "ModelAttention", "ModelIntermediate", "ModelLayer", "ModelOutput", "None", "True", "ValueError", "__init__", "add_cross_attention", "and", "apply_chunking_to_forward", "are", "attention", "attention_mask", "attention_output", "be", ...
tapas/modeling_tapas.py:TapasEncoder
[ -0.00009323761332780123, 0.014523956924676895, 0.019815631210803986, 0.021504463627934456, -0.0001987896248465404, 0.03692913055419922, 0.007374567445367575, -0.041207507252693176, 0.006558298598974943, 0.0099641103297472, 0.011934414505958557, -0.0041657863184809685, -0.0011188513599336147,...
[ "BaseModelOutput", "DynamicCache", "EncoderDecoderCache", "False", "ModelEncoder", "ModelLayer", "Module", "ModuleList", "None", "True", "__init__", "all_attentions", "all_hidden_states", "and", "attention_mask", "attentions", "cache_position", "class", "config", "def", "else...
tapas/modeling_tapas.py:TapasPooler
[ -0.0002915378427132964, 0.011618588119745255, 0.03685896843671799, 0.022893769666552544, -0.001151842763647437, 0.03296703100204468, 0.037087906152009964, -0.0066105760633945465, -0.0011303798528388143, 0.0015954095870256424, 0.014022434130311012, -0.016025640070438385, -0.001244848710484802...
[ "Linear", "ModelPooler", "Module", "Tanh", "__init__", "activation", "class", "config", "def", "dense", "first_token_tensor", "forward", "hidden_size", "hidden_states", "nn", "pooled_output", "return", "self", "super" ]
tapas/modeling_tapas.py:TapasPredictionHeadTransform
[ -0.0002919524849858135, 0.041955187916755676, 0.0442478209733963, 0.03209686279296875, -0.0013899088371545076, 0.026136018335819244, 0.029574967920780182, -0.011348534375429153, -0.0021780014503747225, 0.020519066601991653, 0.01662158966064453, 0.00664863595739007, 0.0001226916938321665, 0...
[ "ACT2FN", "LayerNorm", "Linear", "ModelPredictionHeadTransform", "Module", "__init__", "class", "config", "def", "dense", "else", "eps", "forward", "hidden_act", "hidden_size", "hidden_states", "if", "isinstance", "layer_norm_eps", "nn", "return", "self", "str", "super"...
tapas/modeling_tapas.py:TapasLMPredictionHead
[ -0.00033743318635970354, 0.03515048697590828, 0.03377203643321991, 0.018723953515291214, -0.0015794745413586497, 0.024926980957388878, 0.056746214628219604, -0.029292073100805283, 0.00021807517623528838, 0.004077916033565998, 0.019528048112988472, 0.008155832067131996, -0.0002458954695612192...
[ "Linear", "ModelLMPredictionHead", "ModelPredictionHeadTransform", "Module", "Parameter", "__init__", "bias", "class", "config", "decoder", "def", "forward", "hidden_size", "hidden_states", "nn", "return", "self", "super", "torch", "transform", "vocab_size", "zeros" ]
tapas/modeling_tapas.py:TapasOnlyMLMHead
[ -0.00036667720996774733, 0.01974974200129509, 0.018704168498516083, 0.03159958869218826, -0.0016627541044726968, 0.03694363683462143, 0.04042888432741165, -0.006534841377288103, -0.003775686025619507, -0.006215360015630722, 0.01579979434609413, 0.011210883036255836, -0.002410630462691188, ...
[ "ModelLMPredictionHead", "ModelOnlyMLMHead", "Module", "__init__", "class", "config", "def", "forward", "nn", "prediction_scores", "predictions", "return", "self", "sequence_output", "super" ]
tapas/modeling_tapas.py:TapasPreTrainedModel
[ -0.00027926606708206236, 0.03233082965016365, 0.01144101470708847, 0.006545854266732931, -0.0013376311399042606, 0.022540507838129997, 0.016848459839820862, -0.008025786839425564, -0.004269035533070564, -0.0038705922197550535, 0.00648893415927887, -0.005549746099859476, -0.002106057479977607...
[ "Model", "ModelConfig", "ModelLMPredictionHead", "ModelPreTrainedModel", "PreTrainedModel", "True", "_init_weights", "base_model_prefix", "bias", "class", "config", "def", "if", "init", "isinstance", "module", "no_grad", "self", "super", "supports_gradient_checkpointing", "to...
tapas/modeling_tapas.py:TapasModel
[ 0.00004267695840098895, 0.021738547831773758, 0.007171479519456625, 0.026332776993513107, 0.0001829637767514214, 0.03787437826395035, 0.006779289338737726, 0, 0.007731751538813114, -0.0033476243261247873, 0.02050594985485077, -0.003389644669368863, 0.00048323447117581964, 0.001834890339523...
[ "BaseModelOutputWithPooling", "ModelEmbeddings", "ModelEncoder", "ModelModel", "ModelPooler", "ModelPreTrainedModel", "None", "True", "_", "__init__", "add_pooling_layer", "and", "attention_mask", "attentions", "auto_docstring", "class", "config", "def", "device", "dtype", "e...
tapas/modeling_tapas.py:TapasForMaskedLM
[ -0.00009312270412920043, 0.034533727914094925, 0.017712460830807686, -0.009468925185501575, -0.0005326270475052297, 0.043222859501838684, 0.012643800117075443, -0.008856230415403843, 0.002297606784850359, -0.0074080415070056915, 0.02328241616487503, 0.033196937292814255, -0.00100955448579043...
[ "CrossEntropyLoss", "False", "MaskedLMOutput", "Model", "ModelConfig", "ModelForMaskedLM", "ModelModel", "ModelOnlyMLMHead", "ModelPreTrainedModel", "None", "__init__", "_tied_weights_keys", "add_pooling_layer", "attention_mask", "attentions", "auto_docstring", "base_model_prefix", ...
tapas/modeling_tapas.py:TapasForQuestionAnswering
[ -0.00037915314896963537, 0.026820097118616104, -0.018460586667060852, 0.021479297429323196, -0.0015166125958785415, 0.04319080337882042, 0.04249417781829834, 0.014687195420265198, 0, 0.02380138449370861, 0.022988654673099518, 0.008475614711642265, -0.0012190953129902482, 0.0050505376420915...
[ "Bernoulli", "Dropout", "EPSILON_ZERO_DIVISION", "False", "FloatTensor", "IndexMap", "Linear", "LongTensor", "Make", "Model", "ModelForQuestionAnswering", "ModelModel", "ModelPreTrainedModel", "None", "Parameter", "ProductIndexMap", "TableQuestionAnsweringOutput", "True", "ValueE...
tapas/modeling_tapas.py:TapasForSequenceClassification
[ -0.0003463077009655535, 0.02970391884446144, 0.002870426746085286, 0.011767322197556496, -0.001156739192083478, 0.02719051018357277, 0.020107269287109375, 0.012052936479449272, -0.0060835909098386765, 0.010053634643554688, 0.05895085632801056, 0.000642632890958339, -0.00012227875413373113, ...
[ "BCEWithLogitsLoss", "CrossEntropyLoss", "Dropout", "Linear", "MSELoss", "Model", "ModelForSequenceClassification", "ModelModel", "ModelPreTrainedModel", "None", "SequenceClassifierOutput", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "class", "classifier", ...
tapas/modeling_tapas.py:AverageApproximationFunction
[ -0.00024569014203734696, 0.021348141133785248, 0.02651301398873329, 0.042696282267570496, -0.0008572254446335137, 0.06060117483139038, 0.0342029370367527, -0.01463380642235279, 0.005623972974717617, 0.01767534390091896, 0.03649843484163284, -0.021922016516327858, 0.003285433165729046, 0.00...
[ "Enum", "FIRST_ORDER", "ModelApproximationFunction", "RATIO", "SECOND_ORDER", "class", "enum", "first_order", "ratio", "second_order", "str" ]
tapas/modeling_tapas.py:IndexMap
[ -0.0002970784262288362, -0.021973129361867905, -0.012637396343052387, -0.006034072022885084, -0.0011029258603230119, 0.02516094222664833, 0.005806371103972197, -0.052143488079309464, 0.0044117034412920475, 0.009164958260953426, -0.006802562158554792, 0.01889916881918907, -0.00546481972560286...
[ "ModelMap", "as_tensor", "batch_dims", "batch_shape", "class", "def", "device", "indices", "num_segments", "return", "self", "size", "torch" ]
tapas/modeling_tapas.py:ProductIndexMap
[ -0.0001729882205836475, -0.017534155398607254, 0.004186841659247875, 0.01022825762629509, -0.0004197379166726023, 0.043835390359163284, -0.0043835388496518135, -0.03371953219175339, 0, 0.007867890410125256, -0.01573578082025051, 0.02349127270281315, -0.0020091221667826176, -0.0134878121316...
[ "IndexMap", "ModelIndexMap", "__init__", "batch_dims", "class", "def", "div", "float", "floor", "fmod", "if", "index", "indices", "inner_index", "long", "num_segments", "outer_index", "project_inner", "project_outer", "return", "rounding_mode", "self", "super", "torch",...
tapas/modeling_tapas.py:gather
[ -0.00006930220843059942, -0.0014760929625481367, -0.02113143540918827, 0.0040680840611457825, -0.0000984356229309924, 0.04000282660126686, 0.02079242840409279, -0.022600466385483742, 0.010170210152864456, 0.0013065894600003958, -0.01113072969019413, -0.0036443250719457865, -0.003587824059650...
[ "Model", "batch_dims", "def", "else", "expand", "if", "index", "indices", "len", "name", "return", "segmented_Model", "shape", "size", "torch", "unsqueeze", "values", "view" ]
tapas/modeling_tapas.py:flatten
[ -0.00007161113899201155, -0.010740792378783226, -0.01518331840634346, 0.007366721052676439, -0.00020385011157486588, 0.04003897309303284, -0.0071699004620313644, -0.04138860106468201, 0.0016729767667129636, -0.005145458038896322, -0.012371593154966831, 0.003092898288741708, -0.00517357513308...
[ "IndexMap", "Model", "_", "arange", "batch_dims", "batch_shape", "batch_size", "def", "device", "end", "for", "in", "index", "indices", "len", "list", "name", "num_segments", "offset", "prod", "range", "return", "segmented_Model", "size", "start", "tensor", "torch...
tapas/modeling_tapas.py:range_index_map
[ -0.00016673370555508882, -0.019432375207543373, -0.005812863353639841, 0.0015795824583619833, -0.0003053859400097281, 0.03819079324603081, -0.004240301437675953, -0.05077129229903221, 0.0026256171986460686, 0.006093678064644337, 0.00021675381867680699, -0.009210720658302307, -0.0036225090734...
[ "IndexMap", "Model_index_map", "aModel", "as_tensor", "assert", "batch_dims", "batch_shape", "cat", "cpu", "def", "device", "dim", "dtype", "else", "end", "for", "if", "in", "indices", "int", "is_tensor", "len", "list", "long", "multiples", "name", "new_shape", ...
tapas/modeling_tapas.py:_segment_reduce
[ -0.0002189266961067915, 0.0076271235011518, -0.02723165601491928, 0.0031073465943336487, -0.0006214693421497941, 0.008079101331532001, 0.012937861494719982, -0.01830509677529335, 0.009491532109677792, 0.015706224367022514, 0.0004413844726514071, 0.020677980035543442, -0.003983053378760815, ...
[ "False", "_segment_reduce", "as_tensor", "batch_shape", "cat", "clone", "def", "device", "dim", "dtype", "flat_index", "flat_values", "flatten", "flattened_shape", "float", "include_self", "index", "indices", "int", "len", "long", "name", "new_shape", "num_segments", ...
tapas/modeling_tapas.py:reduce_sum
[ 0.00002255815161333885, -0.008563734591007233, -0.011718794703483582, 0.04777662456035614, 0.00019278966647107154, 0.027156053110957146, 0.019155722111463547, -0.016113342717289925, 0.007662288844585419, -0.005746716633439064, 0.0055213551968336105, 0.001788806403055787, -0.00232403981499373...
[ "Model_sum", "_segment_Model", "def", "index", "name", "return", "segmented_Model_sum", "sum", "values" ]