identifier
stringlengths
24
117
embedding
listlengths
2.56k
2.56k
tokens
listlengths
4
448
conditional_detr/modeling_conditional_detr.py:ConditionalDetrEncoder
[ 0, 0.021749282255768776, 0.010089872404932976, -0.0014013712061569095, -0.00019268854521214962, 0.04439543932676315, -0.000777761044446379, -0.024439914152026176, 0.007231075316667557, -0.004176086280494928, 0.024888353422284126, 0.02466413378715515, -0.000763747317250818, -0.0116033535450...
[ "BaseModelOutput", "ModelEncoder", "ModelEncoderLayer", "ModelPreTrainedModel", "ModelSelfAttention", "ModuleList", "None", "_", "__init__", "_can_record_outputs", "attention_mask", "attentions", "capture_outputs", "class", "config", "create_bidirectional_mask", "def", "dropout", ...
conditional_detr/modeling_conditional_detr.py:gen_sine_position_embeddings
[ -0.00006699279038002715, 0.03385194018483162, 0.008462985046207905, 0.015917137265205383, -0.0002872370823752135, 0.017710616812109947, 0.05290766805410385, 0, 0.014572027139365673, 0.015692953020334244, 0.0034608563873916864, 0.014459935016930103, 0.0006095030694268644, 0.0108729740604758...
[ "Model_sine_position_embeddings", "None", "arange", "cat", "cos", "d_model", "def", "device", "dim", "dim_t", "div", "dtype", "flatten", "float32", "floor", "math", "pi", "pos", "pos_tensor", "pos_x", "pos_y", "return", "rounding_mode", "scale", "sin", "stack", "t...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrDecoder
[ -0.0001746007619658485, 0.05621086433529854, -0.008747674524784088, -0.002031718147918582, -0.0009946952341124415, 0.03250749036669731, 0.0494384728372097, -0.03205599635839462, 0.0007019303739070892, -0.01467351894825697, -0.02234889753162861, 0.03995712101459503, -0.006546646822243929, 0...
[ "LayerNorm", "ModelDecoder", "ModelDecoderCrossAttention", "ModelDecoderLayer", "ModelDecoderOutput", "ModelDecoderSelfAttention", "ModelMLPPredictionHead", "ModelPreTrainedModel", "ModuleList", "None", "OutputRecorder", "_", "__init__", "_can_record_outputs", "and", "attention_mask", ...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrModel
[ -0.00012971900287084281, 0.0379200205206871, 0.0027626643422991037, 0.015818707644939423, -0.0006065240013413131, 0.032310549169778824, 0.020642852410674095, -0.014696813188493252, 0.009031247347593307, 0.007460596039891243, 0.0021456226240843534, 0.013070066459476948, -0.0005469233728945255...
[ "BaseModelOutput", "Conv2d", "Embedding", "False", "ModelConvEncoder", "ModelDecoder", "ModelEncoder", "ModelLearnedPositionEmbedding", "ModelModel", "ModelModelOutput", "ModelPreTrainedModel", "ModelSinePositionEmbedding", "None", "True", "_", "__init__", "attention_mask", "attent...
conditional_detr/modeling_conditional_detr.py:inverse_sigmoid
[ 0.00006263991963351145, 0.023037374019622803, 0.018407313153147697, 0.016374604776501656, 0.00035466544795781374, 0.03929905220866203, 0.036588769406080246, 0.015019464306533337, 0.015132392756640911, 0.013777253217995167, 0.043590325862169266, -0.041105903685092926, 0.0016304024029523134, ...
[ "Model_sigmoid", "clamp", "def", "eps", "log", "max", "min", "return", "torch", "x", "x1", "x2" ]
conditional_detr/modeling_conditional_detr.py:ConditionalDetrForObjectDetection
[ -0.00019663988496176898, 0.019074950367212296, 0.0004655868106056005, 0.004232607316225767, -0.0006489998195320368, 0.03521529212594032, 0.034989554435014725, -0.022799644619226456, 0.007675128057599068, 0.008916692808270454, 0.002384368795901537, 0.03973007574677467, -0.0027935209218412638,...
[ "Linear", "ModelForObjectDetection", "ModelMLPPredictionHead", "ModelModel", "ModelObjectDetectionOutput", "ModelPreTrainedModel", "None", "__init__", "_set_aux_loss", "a", "append", "auto_docstring", "auxiliary_loss", "auxiliary_outputs", "b", "bbox_predictor", "can_return_tuple", ...
conditional_detr/modeling_conditional_detr.py:ConditionalDetrForSegmentation
[ -0.00015192237333394587, 0.0245042871683836, -0.01742277294397354, -0.010790878906846046, -0.0002002213877858594, 0.031698208302259445, 0.023380236700177193, -0.03282225504517555, 0.0035126558504998684, 0.003653162159025669, 0.011128094047307968, 0.031473398208618164, -0.003456453327089548, ...
[ "Model", "ModelForObjectDetection", "ModelForSegmentation", "ModelMHAttentionMap", "ModelMaskHeadSmallConv", "ModelPreTrainedModel", "ModelSegmentationOutput", "None", "__init__", "_checkpoint_conversion_mapping", "activation_function", "adapter1", "adapter2", "adapter3", "attention_mask...
flaubert/modeling_flaubert.py:create_sinusoidal_embeddings
[ -0.00018755829660221934, 0.03315316140651703, 0.028237348422408104, 0.03452501818537712, -0.0008931347401812673, 0.024807710200548172, 0.05578877031803131, -0.010917679406702518, 0.01737682893872261, 0.016576580703258514, 0.018177079036831856, 0.004744331818073988, 0.0004501399234868586, -...
[ "False", "FloatTensor", "Model_sinusoidal_embeddings", "array", "cos", "def", "detach_", "dim", "for", "in", "j", "n_pos", "np", "out", "pos", "position_enc", "power", "range", "requires_grad", "return", "sin", "torch" ]
flaubert/modeling_flaubert.py:get_masks
[ -0.00004481861833482981, -0.003693928709253669, 0.033357296139001846, -0.011865346692502499, 0.00019151761080138385, 0.04119290038943291, -0.0020568466279655695, -0.05731186270713806, 0.019141266122460365, 0.02294713258743286, 0.023618755862116814, 0.018917392939329147, 0.0024906033650040627...
[ "False", "Model_masks", "None", "alen", "arange", "assert", "attn_mask", "bs", "causal", "def", "device", "dtype", "else", "if", "is", "item", "lengths", "long", "mask", "not", "or", "padding_mask", "repeat", "return", "size", "slen", "torch" ]
flaubert/modeling_flaubert.py:MultiHeadAttention
[ -0.00016036011220421642, 0.01905994303524494, 0.03766876459121704, -0.0046240100637078285, -0.000874050660058856, 0.01488705724477768, 0.038796573877334595, -0.024586200714111328, 0.0009092946420423687, 0.03157860413193703, -0.014097591862082481, 0.019172724336385727, -0.0013815639540553093,...
[ "EncoderDecoderCache", "False", "Linear", "ModelHeadAttention", "Module", "None", "True", "__init__", "and", "assert", "attention_dropout", "bs", "cache", "cache_position", "class", "config", "context", "contiguous", "cross_attention_cache", "curr_past_key_values", "current_s...
flaubert/modeling_flaubert.py:TransformerFFN
[ -0.000245455012191087, 0.03669283911585808, 0.02373567968606949, 0.020295726135373116, -0.000670074310619384, 0.017543762922286987, 0.042655423283576965, -0.03715150058269501, 0.004701269790530205, 0.0041566104628145695, 0.014791800640523434, -0.017085103318095207, 0.003841281635686755, -0...
[ "Linear", "ModelFFN", "Module", "__init__", "act", "apply_chunking_to_forward", "chunk_size_feed_forward", "class", "config", "def", "dim_hidden", "dropout", "else", "ff_chunk", "forward", "functional", "gelu", "gelu_activation", "if", "in_dim", "input", "lin1", "lin2", ...
flaubert/modeling_flaubert.py:FlaubertPredLayer
[ -0.00039676710730418563, 0.03571266308426857, 0.017740381881594658, 0.028871601447463036, -0.0015870684292167425, 0.03385746106505394, 0.03779976814985275, -0.007246887777000666, -0.002971224021166563, -0.015305427834391594, 0.024581443518400192, 0.0012247241102159023, -0.0008370155701413751...
[ "AdaptiveLogSoftmaxWithLoss", "False", "Linear", "ModelPredLayer", "Module", "None", "True", "__init__", "asm", "asm_cutoffs", "asm_div_value", "class", "config", "cross_entropy", "cutoffs", "def", "dim", "div_value", "else", "emb_dim", "forward", "functional", "head_bias...
flaubert/modeling_flaubert.py:FlaubertSquadHeadOutput
[ -0.0004275301762390882, 0.014213989488780499, 0.01658298820257187, 0.016701437532901764, -0.001969229895621538, 0.05993565917015076, 0.077229343354702, -0.002872410463169217, 0.008587619289755821, 0.00458993436768651, 0.023453082889318466, 0.03198147565126419, -0.002990860491991043, -0.004...
[ "ModelOutput", "ModelSquadHeadOutput", "None", "class", "cls_logits", "end_top_index", "end_top_log_probs", "loss", "r", "start_top_index", "start_top_log_probs" ]
flaubert/modeling_flaubert.py:FlaubertPoolerStartLogits
[ -0.00026344796060584486, 0.01514019351452589, 0.02305438555777073, 0.027068831026554108, -0.0012903574388474226, 0.04450299218297005, 0.0399150550365448, -0.020531021058559418, 0.0017706571379676461, 0.006652509327977896, 0.019728131592273712, 0.0014982483116909862, 0.00007079044007696211, ...
[ "Linear", "ModelPoolerStartLogits", "Module", "None", "__init__", "class", "config", "def", "dense", "else", "float16", "forward", "hidden_size", "hidden_states", "if", "is", "nn", "not", "p_mask", "return", "self", "squeeze", "super", "torch", "x" ]
flaubert/modeling_flaubert.py:FlaubertPoolerEndLogits
[ -0.0003396792453713715, 0.021855106577277184, 0.03885352239012718, 0.03538445755839348, -0.0014526707818731666, 0.05550503358244896, 0.040934961289167404, -0.029255777597427368, 0.004394148476421833, 0.017229687422513962, 0.03145285323262215, 0.021161293610930443, 0.0020669843070209026, -0...
[ "LayerNorm", "Linear", "ModelPoolerEndLogits", "Module", "None", "One", "Tanh", "__init__", "activation", "assert", "be", "cat", "class", "config", "def", "dense_0", "dense_1", "dim", "else", "eps", "expand", "float16", "forward", "gather", "hidden_size", "hidden_st...
flaubert/modeling_flaubert.py:FlaubertPoolerAnswerClass
[ -0.0003486787900328636, 0.016564039513468742, 0.02576628513634205, 0.03427835926413536, -0.0012437409022822976, 0.046931445598602295, 0.030597463250160217, 0.0007620608666911721, -0.0016391498502343893, 0.01851951703429222, 0.04509099945425987, 0.017484264448285103, 0.0027175378054380417, ...
[ "Linear", "ModelPoolerAnswerClass", "Module", "None", "One", "Tanh", "__init__", "activation", "assert", "be", "cat", "class", "cls_index", "cls_token_state", "config", "def", "dense_0", "dense_1", "dim", "else", "expand", "forward", "gather", "hidden_size", "hidden_s...
flaubert/modeling_flaubert.py:FlaubertSQuADHead
[ -0.00046064614434726536, 0.025353962555527687, 0.027240769937634468, 0.026533218100667, -0.001931028557009995, 0.052594732493162155, 0.04481165483593941, 0.0023290268145501614, -0.002137398114427924, 0.019103916361927986, 0.027830397710204124, 0.026533218100667, 0.0014224753249436617, -0.0...
[ "BCEWithLogitsLoss", "CrossEntropyLoss", "False", "ModelPoolerAnswerClass", "ModelPoolerEndLogits", "ModelPoolerStartLogits", "ModelSQuADHead", "ModelSquadHeadOutput", "Module", "None", "__init__", "and", "answer_class", "auto_docstring", "bh", "bl", "blh", "class", "cls_index", ...
flaubert/modeling_flaubert.py:FlaubertSequenceSummary
[ -0.00037931749830022454, 0.022525623440742493, 0.020658213645219803, 0.04575152322649956, -0.0016193939372897148, 0.025326736271381378, 0.020074648782610893, -0.01207980327308178, -0.0034576246980577707, 0.00682771485298872, 0.02917826734483242, 0.020658213645219803, 0.0016704559093341231, ...
[ "Dropout", "Identity", "Linear", "ModelSequenceSummary", "Module", "None", "NotImplementedError", "__init__", "activation", "activation_string", "and", "attn", "class", "cls_index", "config", "def", "dim", "dtype", "elif", "else", "expand", "first", "first_dropout", "fo...
flaubert/modeling_flaubert.py:FlaubertPreTrainedModel
[ -0.0002607954083941877, 0.05853249132633209, 0.003272446570917964, 0.02663685753941536, -0.0012646791292354465, 0.0281230341643095, 0.027094142511487007, -0.015662014484405518, -0.004572851117700338, -0.0046585919335484505, 0.005973286461085081, 0, -0.0015076118288561702, -0.00194346159696...
[ "Embedding", "False", "LayerNorm", "Linear", "ModelConfig", "ModelModel", "ModelPreTrainedModel", "None", "PreTrainedModel", "_init_weights", "_is_hf_initialized", "and", "arange", "attention_mask", "attns_list", "base_model_prefix", "bias", "class", "config", "copy_", "creat...
flaubert/modeling_flaubert.py:FlaubertModel
[ -0.00006717520591337234, 0.05574927479028702, 0.017309250310063362, 0.008373631164431572, -0.00019230549514759332, 0.053501322865486145, 0.020568784326314926, -0.02135556936264038, 0.006547167897224426, -0.003962019924074411, 0.01719685271382332, 0.010228193365037441, 0.00029855649336241186,...
[ "BaseModelOutput", "DynamicCache", "Embedding", "EncoderDecoderCache", "False", "FloatTensor", "LayerNorm", "ModelModel", "ModelPreTrainedModel", "ModuleList", "MultiHeadAttention", "None", "TransformerFFN", "__init__", "_slen", "a", "and", "append", "arange", "assert", "atte...
flaubert/modeling_flaubert.py:FlaubertWithLMHeadModel
[ -0.00022195327619556338, 0.022434895858168602, 0.006595182698220015, -0.01713620126247406, -0.0007539365324191749, 0.0446443147957325, 0.02390049211680889, -0.01454322412610054, -0.003114391816779971, -0.00873720832169056, 0.013923163525760174, 0.016234295442700386, -0.0020574715454131365, ...
[ "GenerationMixin", "MaskedLMOutput", "ModelModel", "ModelPreTrainedModel", "ModelPredLayer", "ModelWithLMHeadModel", "None", "Tensor", "__init__", "_tied_weights_keys", "attention_mask", "attentions", "auto_docstring", "cache", "cat", "class", "config", "def", "device", "dim", ...
flaubert/modeling_flaubert.py:FlaubertForSequenceClassification
[ -0.0003979328612331301, 0.028233427554368973, 0.014981001615524292, 0.02131911739706993, -0.0012820280389860272, 0.031805820763111115, 0.013770997524261475, -0.001447683316655457, -0.006251687183976173, 0.002477627247571945, 0.046556346118450165, 0.0063669257797300816, -0.0002187730424338951...
[ "BCEWithLogitsLoss", "CrossEntropyLoss", "MSELoss", "ModelForSequenceClassification", "ModelModel", "ModelPreTrainedModel", "ModelSequenceSummary", "None", "SequenceClassifierOutput", "Tensor", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "cache", "class", "...
flaubert/modeling_flaubert.py:FlaubertForTokenClassification
[ -0.00028927106177434325, 0.03725525364279747, 0.009828073903918266, -0.026970062404870987, -0.0010928018018603325, 0.04251212999224663, 0.04182645305991173, 0.0007035357994027436, -0.00214274856261909, 0.004942606668919325, 0.05348300561308861, 0.02331310510635376, -0.0013927866239100695, ...
[ "CrossEntropyLoss", "Dropout", "Linear", "ModelForTokenClassification", "ModelModel", "ModelPreTrainedModel", "None", "Tensor", "TokenClassifierOutput", "__init__", "attention_mask", "attentions", "auto_docstring", "cache", "class", "classifier", "config", "def", "dropout", "el...
flaubert/modeling_flaubert.py:FlaubertForQuestionAnsweringSimple
[ -0.00017516175284981728, 0.030870718881487846, 0.02557537704706192, 0.025800710543990135, -0.0005598134594038129, 0.04979874938726425, 0.03740539774298668, 0.017238028347492218, 0.0004401049227453768, 0.02827938087284565, 0.017238028347492218, 0.018928032368421555, 0.0012322937836870551, 0...
[ "CrossEntropyLoss", "Linear", "ModelForQuestionAnsweringSimple", "ModelModel", "ModelPreTrainedModel", "None", "QuestionAnsweringModelOutput", "Tensor", "__init__", "and", "attention_mask", "attentions", "auto_docstring", "cache", "clamp", "class", "config", "contiguous", "def", ...
flaubert/modeling_flaubert.py:FlaubertForQuestionAnsweringOutput
[ -0.0002433365589240566, 0.02347482182085514, 0.029314899817109108, 0.025077980011701584, -0.0012166828382760286, 0.06229415908455849, 0.07237115502357483, 0.013283313252031803, 0.012023689225316048, 0.004523197188973427, 0.016489630565047264, 0.029200388118624687, -0.001116485451348126, -0...
[ "ModelForQuestionAnsweringOutput", "ModelOutput", "None", "attentions", "class", "cls_logits", "end_top_index", "end_top_log_probs", "hidden_states", "loss", "r", "start_top_index", "start_top_log_probs" ]
flaubert/modeling_flaubert.py:FlaubertForQuestionAnswering
[ -0.0002513910294510424, 0.030138598755002022, 0.016995450481772423, 0.01858169212937355, -0.000810824625659734, 0.047587260603904724, 0.04985332116484642, 0.008271118625998497, 0.007364694960415363, 0.01994132809340954, 0.01812848076224327, 0.01960141956806183, 0.0018128480296581984, -0.00...
[ "ModelForQuestionAnswering", "ModelForQuestionAnsweringOutput", "ModelModel", "ModelPreTrainedModel", "ModelSQuADHead", "None", "Tensor", "__init__", "attention_mask", "attentions", "auto_docstring", "cache", "class", "cls_index", "cls_logits", "config", "def", "else", "end_posit...
flaubert/modeling_flaubert.py:FlaubertForMultipleChoice
[ -0.00026341210468672216, 0.049208227545022964, 0.025629285722970963, 0.03508364409208298, -0.0009183827205561101, 0.04237375408411026, 0.04009559378027916, -0.01879481039941311, -0.004072208888828754, -0.00025451305555179715, 0.030527327209711075, -0.00023671492817811668, -0.0034599534701555...
[ "CrossEntropyLoss", "Linear", "Model", "ModelForMultipleChoice", "ModelModel", "ModelPreTrainedModel", "ModelSequenceSummary", "MultipleChoiceModelOutput", "None", "Please", "Tensor", "The", "__init__", "attention", "attention_mask", "attentions", "auto_docstring", "be", "cache",...
regnet/modeling_regnet.py:RegNetConvLayer
[ -0.000010332522833778057, 0.0037259298842400312, 0.005830653943121433, -0.002830000128597021, 0.00026486811111681163, 0.036406032741069794, 0.0023749247193336487, -0.016837790608406067, 0.02036462537944317, 0.0006968342349864542, 0.04505246505141258, -0.03026251494884491, 0.00308598019182682...
[ "ACT2FN", "BatchNorm2d", "Conv2d", "Identity", "ModelConvLayer", "Module", "None", "__init__", "activation", "class", "convolution", "def", "else", "forward", "groups", "hidden_state", "if", "in_channels", "is", "kernel_size", "nn", "normalization", "not", "out_channels...
regnet/modeling_regnet.py:RegNetEmbeddings
[ -0.000047152152546914294, 0.013650327920913696, 0.01737314462661743, 0.01455282885581255, 0, 0.03203878551721573, 0.019629396498203278, -0.03497191518545151, 0.01235298253595829, 0.009532666765153408, 0.026059716939926147, -0.008912197314202785, 0.0008672470576129854, -0.002369065303355455...
[ "Make", "ModelConvLayer", "ModelEmbeddings", "Module", "ValueError", "__init__", "activation", "channel", "class", "config", "configuration", "def", "dimension", "embedder", "embedding_size", "forward", "hidden_act", "hidden_state", "if", "in", "kernel_size", "match", "nn...
regnet/modeling_regnet.py:RegNetShortCut
[ -0.00010893669241340831, -0.002579054096713662, 0.029928363859653473, 0.006121710874140263, -0.00018598948372527957, 0.031062014400959015, 0.016211196780204773, -0.01995224319398403, 0.014000579714775085, 0.014397356659173965, 0.02448684349656105, -0.031288743019104004, 0.002933319890871644,...
[ "BatchNorm2d", "Conv2d", "ModelShortCut", "Module", "__init__", "class", "convolution", "def", "forward", "hidden_state", "in_channels", "input", "kernel_size", "nn", "normalization", "out_channels", "return", "self", "stride", "super" ]
regnet/modeling_regnet.py:RegNetSELayer
[ -0.00002075294469250366, 0.042728107422590256, 0.010060321539640427, 0.0035324161872267723, 0.00026669743238016963, 0.03888483718037605, 0.022042278200387955, -0.01639041118323803, 0.012999291531741619, 0.009382097981870174, 0.024189986288547516, -0.0012363457353785634, 0.002274875994771719,...
[ "AdaptiveAvgPool2d", "Conv2d", "ModelSELayer", "Module", "ReLU", "Sequential", "Sigmoid", "__init__", "attention", "class", "def", "forward", "hidden_state", "in_channels", "kernel_size", "nn", "pooled", "pooler", "reduced_channels", "return", "self", "super" ]
regnet/modeling_regnet.py:RegNetXLayer
[ -0.00020149024203419685, -0.012553020380437374, 0.017916584387421608, 0.0011982428841292858, -0.0006668792339041829, 0.058428604155778885, 0.0057915071956813335, -0.036517877131700516, 0.006704454310238361, -0.023964857682585716, 0.028529591858386993, -0.03628964349627495, 0.0021967785432934...
[ "ACT2FN", "Identity", "ModelConvLayer", "ModelShortCut", "ModelXLayer", "Module", "None", "Sequential", "__init__", "activation", "class", "config", "def", "else", "forward", "groups", "groups_width", "hidden_act", "hidden_state", "if", "in_channels", "kernel_size", "laye...
regnet/modeling_regnet.py:RegNetYLayer
[ -0.00017524529539514333, -0.0038235336542129517, 0.016993483528494835, 0.004418305587023497, -0.00018763638217933476, 0.05347282811999321, 0.025830093771219254, -0.03783882409334183, 0.00594771932810545, 0, 0.03851856291294098, -0.02684970386326313, 0.0006939005688764155, -0.01189543865621...
[ "ACT2FN", "Identity", "ModelConvLayer", "ModelSELayer", "ModelShortCut", "ModelYLayer", "Module", "None", "Sequential", "__init__", "activation", "class", "config", "def", "else", "forward", "groups", "groups_width", "hidden_act", "hidden_state", "if", "in_channels", "int...
regnet/modeling_regnet.py:RegNetStage
[ -0.00008603272726759315, -0.014304493553936481, -0.0008053373312577605, 0.00516551174223423, 0.00009179781045531854, 0.062213193625211716, 0.022137906402349472, -0.04086998105049133, 0.007322538178414106, 0.011636591516435146, 0.04904397949576378, -0.05562858656048775, 0.0007272869697771966,...
[ "ModelStage", "ModelXLayer", "ModelYLayer", "Module", "Sequential", "_", "__init__", "class", "config", "def", "depth", "else", "for", "forward", "hidden_state", "if", "in", "in_channels", "layer", "layer_type", "layers", "nn", "out_channels", "range", "return", "se...
regnet/modeling_regnet.py:RegNetEncoder
[ -0.00009247014531865716, 0.0033958558924496174, 0.01228708028793335, 0.024235984310507774, 0, 0.036973968148231506, 0.011103462427854538, -0.04463930055499077, 0.0060308147221803665, 0.007778060156852007, 0.03584671393036842, -0.011441639624536037, -0.000979302916675806, 0.0056080939248204...
[ "BaseModelOutputWithNoAttention", "False", "ModelEncoder", "ModelStage", "Module", "ModuleList", "None", "True", "__init__", "append", "class", "config", "def", "depth", "depths", "downsample_in_first_stage", "else", "embedding_size", "for", "forward", "hidden_sizes", "hidd...
regnet/modeling_regnet.py:RegNetPreTrainedModel
[ -0.000033521548175485805, 0.041505370289087296, -0.008637604303658009, 0.01592908799648285, 0.00019192729087080806, 0.01985527202486992, 0.018621329218149185, -0.010264166630804539, 0.007908456027507782, 0.014302526600658894, 0.01873350515961647, -0.02265968918800354, -0.0009464907343499362,...
[ "BatchNorm2d", "Conv2d", "GroupNorm", "Linear", "Model", "ModelConfig", "ModelPreTrainedModel", "ModelYLayer", "None", "PreTrainedModel", "_", "_calculate_fan_in_and_fan_out", "_init_weights", "_no_split_modules", "a", "base_model_prefix", "bias", "bound", "class", "config", ...
regnet/modeling_regnet.py:RegNetModel
[ 0.00004328732393332757, 0.025745024904608727, 0.010969619266688824, 0.03716238588094711, 0.0003707843425218016, 0.028655333444476128, 0.022386979311704636, -0.01555895060300827, 0.01304041501134634, 0.030222421512007713, 0.026752440258860588, -0.003008250379934907, 0.0018749095033854246, -...
[ "AdaptiveAvgPool2d", "BaseModelOutputWithPoolingAndNoAttention", "ModelEmbeddings", "ModelEncoder", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "auto_docstring", "class", "config", "def", "else", "embedder", "embedding_output", "encoder", "encoder_outputs", "forward",...
regnet/modeling_regnet.py:RegNetForImageClassification
[ -0.000042190989915980026, 0.031177837401628494, 0.008740929886698723, 0.027837354689836502, 0.00009003644663607702, 0.006207730155438185, 0.02171313762664795, -0.008295531384646893, 0.0046488381922245026, 0.00823985692113638, 0.047212153673172, 0.009631725028157234, 0.000953429436776787, -...
[ "Flatten", "Identity", "ImageClassifierOutputWithNoAttention", "Linear", "Model", "ModelForImageClassification", "ModelModel", "ModelPreTrainedModel", "None", "Sequential", "__init__", "auto_docstring", "class", "classifier", "config", "def", "else", "forward", "hidden_sizes", ...
glm4_moe/modeling_glm4_moe.py:Glm4MoeRotaryEmbedding
[ -0.00030098692514002323, 0.04780977591872215, 0.004612714983522892, -0.006150286644697189, -0.0016971309669315815, 0.03643754869699478, 0.03875841200351715, 0.012590681202709675, -0.004351617768406868, 0.02146798186004162, 0.002509433077648282, -0.0033362405374646187, -0.0013707596808671951,...
[ "False", "ModelRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "arange", "attention_factor", "attention_scaling", "base", "cat", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cpu", "def", "default", "device", ...
glm4_moe/modeling_glm4_moe.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
glm4_moe/modeling_glm4_moe.py:eager_attention_forward
[ 0, 0.020139193162322044, 0.013972979038953781, -0.018215786665678024, 0, 0.038468122482299805, 0.0527239553630352, -0.03326360881328583, 0.020704900845885277, 0.007297629024833441, 0.0280591007322073, 0.020704900845885277, 0.002489113714545965, -0.017876362428069115, -0.03484759107232094...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "key_states", "kwargs", "matmul", "module", "nn", "not", "num_key_value_groups", "p", "query",...
glm4_moe/modeling_glm4_moe.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
glm4_moe/modeling_glm4_moe.py:apply_rotary_pos_emb
[ -0.00020313466666266322, 0.028396088629961014, 0.021097494289278984, -0.006728390231728554, -0.0008695589494891465, 0.02337830513715744, 0.04196690768003464, -0.0029650533106178045, 0.011575112119317055, 0.03329982981085777, 0.0030790939927101135, 0.00803985632956028, -0.0009693444008007646,...
[ "Model_rotary_pos_emb", "cat", "cos", "def", "dim", "k", "k_embed", "k_pass", "k_rot", "q", "q_embed", "q_pass", "q_rot", "return", "rotary_dim", "rotate_half", "shape", "sin", "torch", "unsqueeze", "unsqueeze_dim" ]
glm4_moe/modeling_glm4_moe.py:Glm4MoeAttention
[ -0.000058884113968815655, 0.038698285818099976, 0.019349142909049988, 0.002812375547364354, -0.00020126062736380845, 0.0247489046305418, 0.035098448395729065, -0.013836887665092945, 0.00023729418171569705, 0.02441141940653324, 0.018224192783236504, 0.03127361461520195, 0.0004499800852499902,...
[ "ALL_ATTENTION_FUNCTIONS", "Linear", "ModelAttention", "ModelRMSNorm", "Module", "None", "Tensor", "True", "__init__", "_attn_implementation", "apply_rotary_pos_emb", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "cache_kwargs", "cache...
glm4_moe/modeling_glm4_moe.py:Glm4MoeMLP
[ -0.00024263472005259246, 0.02898676134645939, 0.02358049899339676, 0.02829659916460514, -0.0008483228739351034, 0.05774346739053726, 0.029906975105404854, -0.005061180330812931, -0.001998591236770153, -0.004831126891076565, 0.028066545724868774, -0.04946153610944748, -0.0024299416691064835, ...
[ "ACT2FN", "Linear", "ModelMLP", "Module", "None", "__init__", "act_fn", "class", "config", "def", "down_proj", "else", "forward", "gate_proj", "hidden_act", "hidden_size", "if", "intermediate_size", "is", "nn", "return", "self", "super", "up_proj", "x" ]
glm4_moe/modeling_glm4_moe.py:Glm4MoeTopkRouter
[ -0.00039330136496573687, 0.03474346548318863, -0.0025112200528383255, -0.00980853009968996, -0.001624907017685473, 0.06617803126573563, 0.07326853275299072, -0.030134640634059906, -0.008390429429709911, 0.010576668195426464, 0.021980561316013336, -0.020916985347867012, -0.0009823301807045937...
[ "F", "ModelTopkRouter", "Module", "Parameter", "__init__", "class", "config", "def", "dtype", "e_score_correction_bias", "empty", "float32", "forward", "hidden_size", "hidden_states", "linear", "n_group", "n_routed_experts", "nn", "norm_topk_prob", "num_experts_per_tok", "r...
glm4_moe/modeling_glm4_moe.py:Glm4MoeRMSNorm
[ -0.00009835156379267573, 0.04154934734106064, 0.0313878208398819, 0.048549506813287735, -0.00046573654981330037, 0.03861379250884056, 0.02461347170174122, -0.03003295138478279, 0.008129219524562359, 0.04358164966106415, 0.01964561454951763, 0.004770271480083466, 0.0023992490023374557, 0.01...
[ "ModelRMSNorm", "Module", "Parameter", "True", "__init__", "class", "def", "dtype", "eps", "extra_repr", "f", "float32", "forward", "hidden_size", "hidden_states", "input_dtype", "keepdim", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "shape", "super", ...
glm4_moe/modeling_glm4_moe.py:Glm4MoeNaiveMoe
[ -0.00035851384745910764, 0.041463855654001236, 0.00232943007722497, -0.0018635441083461046, -0.0012811865890398622, 0.054275721311569214, 0.05334394797682762, -0.01584012433886528, -0.005095628555864096, -0.012753630056977272, 0.014093051664531231, -0.030515534803271294, -0.00012921057350467...
[ "ACT2FN", "ModelNaiveMoe", "Module", "None", "Parameter", "__init__", "act_fn", "chunk", "class", "config", "continue", "current_hidden_states", "current_state", "def", "dim", "down_proj", "dtype", "empty", "expert_hit", "expert_idx", "expert_mask", "final_hidden_states", ...
glm4_moe/modeling_glm4_moe.py:Glm4MoeMoE
[ -0.0004175463109277189, 0.03760134056210518, -0.012415536679327488, -0.012297293171286583, -0.001625844044610858, 0.06952700763940811, 0.053445927798748016, -0.025777019560337067, -0.007922294549643993, -0.020810803398489952, 0.030743233859539032, -0.02885134145617485, -0.0007907514227554202...
[ "False", "ModelMLP", "ModelMoE", "ModelNaiveMoe", "ModelTopkRouter", "Module", "True", "__init__", "bool", "class", "config", "def", "denominator", "dim", "e_score_correction_bias", "expand", "experts", "forward", "gate", "gather", "group_idx", "group_mask", "group_scores...
glm4_moe/modeling_glm4_moe.py:Glm4MoeDecoderLayer
[ -0.0002400130615569651, 0.04778926819562912, 0.0105819096788764, -0.0029299373272806406, -0.0007822648040018976, 0.035500600934028625, 0.041872501373291016, -0.03686600551009178, 0.004835818894207478, -0.007139944471418858, -0.0037548711989074945, 0.015929756686091423, -0.0013085156679153442...
[ "False", "GradientCheckpointingLayer", "ModelAttention", "ModelDecoderLayer", "ModelMLP", "ModelMoE", "ModelRMSNorm", "None", "Tensor", "_", "__init__", "attention_mask", "cache_position", "class", "config", "def", "else", "eps", "first_k_dense_replace", "forward", "hidden_si...
glm4_moe/modeling_glm4_moe.py:Glm4MoePreTrainedModel
[ -0.00037101644556969404, 0.05074922740459442, 0.0014767908724024892, -0.005674369167536497, -0.0014840657822787762, 0.041670236736536026, 0.03398801386356354, -0.018972761929035187, -0.005121481604874134, -0.0031136281322687864, 0.008613401092588902, -0.009660976938903332, -0.002226098673418...
[ "ModelAttention", "ModelConfig", "ModelDecoderLayer", "ModelNaiveMoe", "ModelPreTrainedModel", "ModelTopkRouter", "PreTrainedModel", "True", "_can_compile_fullgraph", "_can_record_outputs", "_init_weights", "_keep_in_fp32_modules_strict", "_no_split_modules", "_skip_keys_device_placement",...
glm4_moe/modeling_glm4_moe.py:Glm4MoeModel
[ -0.00012942355533596128, 0.04868438467383385, 0.0003750641772057861, -0.004733204375952482, -0.0006022157031111419, 0.03673867881298065, 0.029751569032669067, -0.016904300078749657, 0.009410060942173004, 0.009804494678974152, 0.010593362152576447, 0.004902247339487076, -0.001556604402139783,...
[ "BaseModelOutputWithPast", "DynamicCache", "Embedding", "False", "ModelDecoderLayer", "ModelModel", "ModelPreTrainedModel", "ModelRMSNorm", "ModelRotaryEmbedding", "ModuleList", "None", "ValueError", "You", "__init__", "_keys_to_ignore_on_load_unexpected", "and", "arange", "attenti...
glm4_moe/modeling_glm4_moe.py:Glm4MoeForCausalLM
[ -0.00028262686100788414, 0.03526614233851433, 0.008361488580703735, -0.001301150070503354, -0.0012087186332792044, 0.02741658128798008, 0.0389065183699131, -0.007394513580948114, 0.003640376031398773, 0.026734011247754097, 0.025823917239904404, 0.0026734010316431522, 0.0009385344455949962, ...
[ "CausalLMOutputWithPast", "GenerationMixin", "Linear", "ModelForCausalLM", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "_pp_plan", "_tied_weights_keys", "_tp_plan", "attention_mask", "attentions", "auto_docstring", "cache_position", "can_return_tuple", "class", "colwi...
swin/modeling_swin.py:SwinEncoderOutput
[ -0.00021715721231885254, 0.02585153840482235, 0.03431620076298714, 0.020132170990109444, -0.0012082161847501993, 0.038434144109487534, 0.03729027137160301, -0.05284694954752922, 0.01509912870824337, 0.017615649849176407, -0.004146541003137827, 0.035917624831199646, -0.00303126429207623, 0....
[ "ModelEncoderOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "last_hidden_state", "r", "reshaped_hidden_states" ]
swin/modeling_swin.py:SwinModelOutput
[ -0.00020244609913788736, 0.025114018470048904, 0.02728295512497425, 0.01712319441139698, -0.0010916036553680897, 0.04520523175597191, 0.032648224383592606, -0.023059234023094177, 0.014954255893826485, 0.0061928885988891125, -0.00038705553743056953, 0.014954255893826485, -0.001319912844337523...
[ "ModelModelOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "last_hidden_state", "pooler_output", "r", "reshaped_hidden_states" ]
swin/modeling_swin.py:SwinMaskedImageModelingOutput
[ -0.00019999424694105983, 0.039871420711278915, 0.03443440794944763, -0.008155517280101776, -0.0010760752484202385, 0.05097198486328125, 0.011610285378992558, -0.04938618838787079, 0.015178324654698372, -0.00662635825574398, -0.012233276851475239, 0.03103627637028694, -0.0024070104118436575, ...
[ "ModelMaskedImageModelingOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "loss", "r", "reconstruction", "reshaped_hidden_states" ]
swin/modeling_swin.py:SwinImageClassifierOutput
[ -0.0002069334004772827, 0.034184690564870834, 0.02705344557762146, 0.017092345282435417, -0.0009692265302874148, 0.03327913582324982, 0.04437217861413956, -0.018903454765677452, 0.013470126315951347, -0.006197390146553516, 0.022638866677880287, 0.018450677394866943, -0.003339232876896858, ...
[ "ModelImageClassifierOutput", "ModelOutput", "None", "attentions", "class", "hidden_states", "logits", "loss", "r", "reshaped_hidden_states" ]
swin/modeling_swin.py:window_partition
[ -0.00004125074337935075, 0.009773793630301952, -0.008818882517516613, 0.00499923899769783, 0, 0.025501737371087074, 0.019772272557020187, -0.03999391198158264, 0.012245327234268188, 0.007920143194496632, -0.00150258035864681, -0.04403823986649513, 0.002359191421419382, -0.00117959571070969...
[ "Model_partition", "Model_size", "Models", "batch_size", "contiguous", "def", "height", "input_feature", "num_channels", "permute", "return", "shape", "view", "width" ]
swin/modeling_swin.py:window_reverse
[ -0.0000877177226357162, 0.022455736994743347, 0.02098877914249897, 0.02098877914249897, -0.0002803440729621798, 0.03836658596992493, 0.01602369174361229, -0.043783046305179596, 0.017152121290564537, 0.027646509930491447, -0.017942020669579506, -0.0253896526992321, -0.0025530706625431776, 0...
[ "Model_reverse", "Model_size", "Models", "contiguous", "def", "height", "num_channels", "permute", "return", "shape", "view", "width" ]
swin/modeling_swin.py:SwinEmbeddings
[ -0.00022775810793973505, 0.02710321545600891, 0.011615663766860962, 0.005551604088395834, -0.0011316731106489897, 0.017992891371250153, 0.023231327533721924, -0.026761578395962715, 0.0014448404544964433, 0.005864771082997322, 0.02175089903175831, 0.030747344717383385, -0.002405694918707013, ...
[ "Dropout", "False", "LayerNorm", "ModelEmbeddings", "ModelPatchEmbeddings", "Module", "None", "Parameter", "_", "__init__", "align_corners", "and", "batch_size", "bicubic", "bool_masked_pos", "cat", "class", "class_pos_embed", "config", "def", "dim", "dropout", "else", ...
swin/modeling_swin.py:SwinPatchEmbeddings
[ -0.00009367455641040578, 0.0035018527414649725, 0.020058611407876015, 0.009356950409710407, -0.0003029102517757565, -0.0034598305355757475, 0.014791825786232948, -0.020506849512457848, 0.004006119444966316, 0.0035018527414649725, 0.007732090540230274, 0.010757691226899624, -0.002395267132669...
[ "Conv2d", "Iterable", "ModelPatchEmbeddings", "Module", "_", "__init__", "abc", "class", "collections", "config", "def", "else", "embed_dim", "embeddings", "flatten", "forward", "functional", "grid_size", "height", "hidden_size", "if", "image_size", "isinstance", "kerne...
swin/modeling_swin.py:SwinPatchMerging
[ -0.00021154274872969836, 0.02943691983819008, 0.017414972186088562, 0.015842005610466003, -0.00020013173343613744, 0.013257849030196667, -0.0029633541125804186, -0.03865000233054161, 0.0009409702615812421, 0.011347820051014423, 0.005112136714160442, -0.0016291425563395023, -0.000811060192063...
[ "LayerNorm", "Linear", "ModelPatchMerging", "Module", "__init__", "batch_size", "cat", "class", "def", "dim", "forward", "functional", "height", "if", "input_dimensions", "input_feature", "input_feature_0", "input_feature_1", "input_feature_2", "input_feature_3", "input_resol...
swin/modeling_swin.py:drop_path
[ 0, 0.017753854393959045, 0.03775503486394882, -0.02123720571398735, 0, 0.043598074465990067, 0.0339345820248127, -0.027304979041218758, 0.014832334592938423, 0.0017486985307186842, 0.020113544538617134, -0.04944111406803131, 0.0016574009787291288, -0.018315685912966728, -0.05034004524350...
[ "False", "Model_path", "Model_prob", "def", "device", "div", "dtype", "floor_", "if", "input", "keep_prob", "ndim", "not", "or", "output", "rand", "random_tensor", "return", "shape", "torch", "training" ]
swin/modeling_swin.py:SwinDropPath
[ -0.00011107319733127952, 0.018094833940267563, 0.03596488758921623, -0.025063030421733856, -0.00047414645086973906, 0.031918834894895554, 0.033717080950737, -0.026861274614930153, 0.0066872211173176765, -0.002430439693853259, 0.009721758775413036, -0.033717080950737, -0.000818341679405421, ...
[ "ModelDropPath", "Module", "None", "__init__", "class", "def", "drop_path", "drop_prob", "extra_repr", "f", "forward", "hidden_states", "nn", "p", "return", "self", "super", "training" ]
swin/modeling_swin.py:SwinSelfAttention
[ 0.000012227195838931948, 0.03577328100800514, 0.0438222698867321, -0.010676088742911816, -0.00010436784941703081, 0.009893547743558884, 0.022246509790420532, -0.013191397301852703, 0.005645471159368753, 0.025712044909596443, 0.016321560367941856, 0.031525205820798874, 0.001355472020804882, ...
[ "Dropout", "False", "Iterable", "Linear", "ModelSelfAttention", "Module", "None", "Parameter", "The", "ValueError", "__init__", "a", "abc", "all_head_size", "arange", "attention", "attention_head_size", "attention_mask", "attention_probs", "attention_probs_dropout_prob", "att...
swin/modeling_swin.py:SwinSelfOutput
[ -0.00004677825199905783, 0.04611024260520935, 0.04879627376794815, 0.0014549348270520568, -0.00040045680361799896, 0.04051433876156807, 0.03379925340414047, -0.024510055780410767, 0.0021124533377587795, 0.014885102398693562, 0.010744133964180946, 0.011583519168198109, 0.002350279362872243, ...
[ "Dropout", "Linear", "ModelSelfOutput", "Module", "__init__", "attention_probs_dropout_prob", "class", "config", "def", "dense", "dim", "dropout", "forward", "hidden_states", "input_tensor", "nn", "return", "self", "super" ]
swin/modeling_swin.py:SwinAttention
[ 0.00015754895866848528, 0.031094355508685112, 0.04551492631435394, 0.0031685824505984783, 0.0005386589909903705, 0.03650207072496414, 0.04461364075541496, -0.018589017912745476, 0.014758553355932236, 0.0018589016981422901, 0.015096534974873066, 0.02320810593664646, 0.0020138102117925882, -...
[ "False", "ModelAttention", "ModelSelfAttention", "ModelSelfOutput", "Module", "None", "__init__", "attention_mask", "attention_output", "class", "config", "def", "dim", "forward", "hidden_states", "nn", "num_heads", "output", "output_attentions", "outputs", "return", "self"...
swin/modeling_swin.py:SwinIntermediate
[ -0.0002936642267741263, 0.023951541632413864, 0.039881035685539246, 0.0161586944013834, -0.0009311304893344641, 0.04263145104050636, 0.028764769434928894, -0.018565310165286064, -0.004383475985378027, -0.002549865050241351, 0.02097192406654358, -0.025326749309897423, -0.0019768616184592247, ...
[ "ACT2FN", "Linear", "ModelIntermediate", "Module", "__init__", "class", "config", "def", "dense", "dim", "else", "forward", "hidden_act", "hidden_states", "if", "int", "intermediate_act_fn", "isinstance", "mlp_ratio", "nn", "return", "self", "str", "super" ]
swin/modeling_swin.py:SwinOutput
[ -0.00027769507141783834, 0.044259220361709595, 0.04999227821826935, 0.024652156978845596, -0.0011824435787275434, 0.03600361570715904, 0.04609379917383194, -0.021097660064697266, -0.0009244558750651777, 0.00905823428183794, 0.007968952879309654, -0.007309651002287865, -0.00014690964599139988...
[ "Dropout", "Linear", "ModelOutput", "Module", "__init__", "class", "config", "def", "dense", "dim", "dropout", "forward", "hidden_dropout_prob", "hidden_states", "int", "mlp_ratio", "nn", "return", "self", "super" ]
swin/modeling_swin.py:SwinLayer
[ -0.0001743055909173563, 0.02107160910964012, 0.021296974271535873, -0.010761169716715813, -0.000531720113940537, 0.04304467886686325, 0.022085750475525856, -0.035833004862070084, 0.002000112785026431, 0.01994478516280651, 0.01498675998300314, -0.013352865353226662, 0.0005422840476967394, 0...
[ "False", "Identity", "LayerNorm", "ModelAttention", "ModelDropPath", "ModelIntermediate", "ModelLayer", "ModelOutput", "Module", "None", "_", "__init__", "always_partition", "attention", "attention_output", "attention_outputs", "attention_windows", "attn_mask", "batch_size", "c...
swin/modeling_swin.py:SwinStage
[ -0.00017466909775976092, -0.002935852389782667, 0.0011080019176006317, 0.009710895828902721, -0.00009924380719894543, 0.037714410573244095, 0.0231480672955513, -0.027326010167598724, 0.001164460671134293, 0.014171133749186993, 0.022809313610196114, -0.0287939365953207, -0.001277378061786294,...
[ "False", "GradientCheckpointingLayer", "LayerNorm", "ModelLayer", "ModelStage", "ModuleList", "None", "__init__", "always_partition", "blocks", "class", "config", "def", "depth", "dim", "downsample", "drop_path", "drop_path_rate", "else", "enumerate", "for", "forward", "h...
swin/modeling_swin.py:SwinEncoder
[ -0.00025126183754764497, 0.022125298157334328, 0.018475765362381935, 0.024634353816509247, -0.0008232835098169744, 0.02634507231414318, 0.011975033208727837, -0.04721584543585777, 0.005046620965003967, 0.026002928614616394, 0.003592509776353836, -0.015966709703207016, -0.0022667027078568935,...
[ "False", "ModelEncoder", "ModelEncoderOutput", "ModelPatchMerging", "ModelStage", "Module", "ModuleList", "None", "True", "_", "__init__", "all_hidden_states", "all_reshaped_hidden_states", "all_self_attentions", "always_partition", "and", "attentions", "batch_size", "class", "...
swin/modeling_swin.py:SwinPreTrainedModel
[ -0.00024896516697481275, 0.03414379432797432, -0.006800305563956499, 0.012121045961976051, -0.0012163725914433599, 0.01912052370607853, 0.016047582030296326, -0.0023331590928137302, -0.0007148856529965997, 0.019461961463093758, 0.004524052608758211, 0.007739259861409664, -0.00217666686512529...
[ "Model", "ModelConfig", "ModelEmbeddings", "ModelPreTrainedModel", "ModelSelfAttention", "ModelStage", "None", "PreTrainedModel", "True", "_init_weights", "_no_split_modules", "base_model_prefix", "class", "config", "copy_", "create_relative_position_index", "def", "elif", "if", ...
swin/modeling_swin.py:SwinModel
[ -0.000011052514309994876, 0.036978866904973984, 0.0013026646338403225, 0.026781665161252022, 0.00009454824612475932, 0.036530639976263046, 0.014679490588605404, -0.01378303300589323, 0.008684431202709675, 0.016472404822707176, 0.024988749995827675, -0.001099561108276248, -0.00018734559125732...
[ "AdaptiveAvgPool1d", "False", "LayerNorm", "ModelEmbeddings", "ModelEncoder", "ModelModel", "ModelModelOutput", "ModelPreTrainedModel", "None", "True", "__init__", "add_pooling_layer", "attentions", "auto_docstring", "bool_masked_pos", "class", "config", "def", "depths", "else"...
swin/modeling_swin.py:SwinForMaskedImageModeling
[ -0.0001356861466774717, 0.05692164972424507, 0.0139502864331007, -0.0011275081196799874, -0.0006793061038479209, 0.03383924812078476, 0.0061347647570073605, -0.060059066861867905, 0.010756847448647022, 0.012101453728973866, 0.013334008865058422, 0.019720887765288353, -0.001281577511690557, ...
[ "Conv2d", "False", "Model", "ModelForMaskedImageModeling", "ModelMaskedImageModelingOutput", "ModelModel", "ModelPreTrainedModel", "None", "PixelShuffle", "Sequential", "True", "__init__", "add_pooling_layer", "attentions", "auto_docstring", "batch_size", "bool_masked_pos", "class"...
swin/modeling_swin.py:SwinForImageClassification
[ -0.00009624342783354223, 0.038351044058799744, 0.012876586057245731, 0.022631576284766197, -0.0001829060638556257, 0.00443155225366354, 0.026422087103128433, -0.006298936437815428, 0.003219146514311433, 0.010925588198006153, 0.04370235279202461, 0.00468239514157176, 0.00018116409773938358, ...
[ "False", "Identity", "Linear", "Model", "ModelForImageClassification", "ModelImageClassifierOutput", "ModelModel", "ModelPreTrainedModel", "None", "__init__", "attentions", "auto_docstring", "class", "classifier", "config", "def", "else", "forward", "hidden_states", "if", "in...
swin/modeling_swin.py:SwinBackbone
[ -0.00023546999727841467, 0.0026485947892069817, 0.028553834185004234, 0.04441707581281662, -0.0007081804214976728, 0.015863241627812386, 0.006515259854495525, -0.009291327558457851, -0.0033284479286521673, 0.005523807369172573, 0.008384856395423412, 0.0032859572675079107, -0.0017987783066928...
[ "BackboneMixin", "BackboneOutput", "LayerNorm", "ModelBackbone", "ModelEmbeddings", "ModelEncoder", "ModelPreTrainedModel", "ModuleDict", "None", "True", "__init__", "always_partition", "attentions", "batch_size", "channels", "class", "config", "contiguous", "def", "depths", ...
qwen3_5_moe/modeling_qwen3_5_moe.py:Qwen3_5MoeVisionRotaryEmbedding
[ -0.00027588062221184373, 0.02608063630759716, 0.016156146302819252, 0.009982191026210785, -0.0012261361116543412, 0.045929618179798126, 0.06416298449039459, -0.016617750748991966, 0.004875694401562214, 0.033927906304597855, 0.013098019175231457, -0.018579568713903427, -0.002481122501194477, ...
[ "False", "Model_5MoeVisionRotaryEmbedding", "Module", "Tensor", "__init__", "arange", "class", "def", "device", "dim", "dtype", "float", "forward", "freqs", "inv_freq", "nn", "outer", "persistent", "register_buffer", "return", "self", "seq", "seqlen", "super", "theta"...
qwen3_5_moe/modeling_qwen3_5_moe.py:Qwen3_5MoeTextRotaryEmbedding
[ -0.0002973411465063691, 0.03550688549876213, 0.005511689465492964, -0.008586632087826729, -0.0013199045788496733, 0.05198393762111664, 0.0317937470972538, -0.014504446648061275, -0.008238526061177254, 0.02390332706272602, 0.019261904060840607, -0.0022046759258955717, -0.0011168423807248473, ...
[ "False", "Model_5MoeTextRotaryEmbedding", "Module", "None", "ROPE_INIT_FUNCTIONS", "Tensor", "__init__", "and", "apply_interleaved_mrope", "arange", "attention_factor", "attention_scaling", "base", "cat", "class", "clone", "compute_default_rope_parameters", "config", "cos", "cp...
qwen3_5_moe/modeling_qwen3_5_moe.py:Qwen3_5MoeDynamicCache
[ -0.00019375044212210923, -0.006904561538249254, 0.022432779893279076, -0.008680020458996296, -0.0005143193411640823, 0.019501863047480583, 0.0031845527701079845, -0.05997104570269585, -0.0026631879154592752, 0.03629826381802559, -0.0032972802873700857, 0.01003275066614151, 0.0013950031716376...
[ "Any", "False", "Model_5MoeDynamicCache", "None", "_", "__init__", "__len__", "beam_idx", "cache_kwargs", "cache_position", "cat", "class", "config", "conv_states", "def", "device", "dim", "else", "for", "full_attention", "get_mask_sizes", "get_seq_length", "has_previous_...
qwen3_5_moe/modeling_qwen3_5_moe.py:Qwen3_5MoeRMSNormGated
[ -0.0002094664960168302, 0.014551655389368534, -0.0023488893639296293, 0.049269385635852814, -0.0008951560012064874, 0.029332470148801804, 0.029561631381511688, -0.03276987001299858, 0.0053566135466098785, 0.03231155127286911, 0.02028065361082554, -0.03001995012164116, 0.0007519310456700623, ...
[ "F", "Model_5MoeRMSNormGated", "Module", "None", "Parameter", "True", "__init__", "class", "def", "eps", "float32", "forward", "gate", "hidden_size", "hidden_states", "keepdim", "kwargs", "mean", "nn", "ones", "pow", "return", "rsqrt", "self", "silu", "super", "to...
qwen3_5_moe/modeling_qwen3_5_moe.py:apply_mask_to_padding_states
[ 0.00005794858589069918, -0.0029697560239583254, 0.05510082468390465, -0.006776063237339258, 0.00009846902685239911, -0.00038864766247570515, -0.003457744140177965, -0.06960104405879974, 0.019296443089842796, 0.0010666025336831808, 0.03480052202939987, 0.028777355328202248, 0.0040712147019803...
[ "Model_mask_to_padding_states", "None", "and", "attention_mask", "def", "hidden_states", "if", "is", "not", "return", "shape" ]
qwen3_5_moe/modeling_qwen3_5_moe.py:torch_causal_conv1d_update
[ -0.00022366855409927666, 0.038402725011110306, 0.020926034078001976, -0.015981972217559814, -0.0007293930393643677, 0.026559967547655106, 0.0027019879780709743, -0.061628323048353195, 0.01885642670094967, 0.019891230389475822, 0.04507146030664444, -0.013107516802847385, 0.004455405753105879,...
[ "F", "Model", "Model_causal_conv1d_update", "None", "_", "activation", "bias", "cat", "conv1d", "conv_state", "copy_", "def", "dim", "dtype", "groups", "hidden_size", "hidden_states", "hidden_states_new", "out", "padding", "return", "seq_len", "shape", "silu", "state_...
qwen3_5_moe/modeling_qwen3_5_moe.py:l2norm
[ -0.00002199482150899712, 0.033167753368616104, 0.03361596539616585, 0.03361596539616585, 0.0001873389701358974, 0.035857029259204865, 0.00885220430791378, 0.011485454626381397, 0.011709561571478844, 0.040115050971508026, 0.014566917903721333, -0.005014381371438503, 0.0015967583749443293, 0...
[ "Model", "True", "def", "dim", "eps", "inv_norm", "keepdim", "return", "rsqrt", "sum", "torch", "x" ]
qwen3_5_moe/modeling_qwen3_5_moe.py:torch_chunk_gated_delta_rule
[ -0.0003130299155600369, -0.00805962085723877, 0.017846303060650826, -0.014392179436981678, -0.0009282956016249955, 0.03177793323993683, 0.024869685992598534, -0.036383431404829025, 0.007483933120965958, 0.038686178624629974, 0.02855408377945423, -0.021645838394761086, 0.002446670550853014, ...
[ "F", "False", "Model", "Model_chunk_gated_delta_rule", "None", "attn", "attn_inter", "batch_size", "beta", "bool", "chunk_size", "clone", "contiguous", "core_attn_out", "cumsum", "decay_mask", "def", "device", "diagonal", "dim", "dtype", "else", "eps", "exp", "eye", ...
qwen3_5_moe/modeling_qwen3_5_moe.py:torch_recurrent_gated_delta_rule
[ -0.00026922550750896335, 0.009185760281980038, 0.0362866036593914, -0.0130654601380229, -0.0006953506381250918, 0.04587174579501152, 0.010383903048932552, -0.033776212483644485, 0.0009413977386429906, 0.029440075159072876, 0.0451870933175087, -0.023620525375008583, 0.0017829503631219268, -...
[ "False", "Model", "Model_recurrent_gated_delta_rule", "None", "batch_size", "beta", "beta_t", "contiguous", "core_attn_out", "def", "delta", "dim", "else", "eps", "exp", "float32", "for", "g", "g_t", "i", "if", "in", "initial_state", "is", "k_head_dim", "k_t", "ke...
qwen3_5_moe/modeling_qwen3_5_moe.py:Qwen3_5MoeGatedDeltaNet
[ -0.0002509625337552279, 0.007084313780069351, 0.007227721158415079, -0.0032266611233353615, -0.0009608279797248542, 0.03579442948102951, 0.03143484890460968, -0.03946565091609955, 0.005047932267189026, 0.015258521772921085, 0.018929746001958847, -0.0309759471565485, 0.00125481269787997, 0....
[ "A", "ACT2FN", "A_log", "Conv1d", "F", "Falling", "FusedRMSNormGated", "Linear", "Model_5MoeGatedDeltaNet", "Model_5MoeRMSNormGated", "Module", "None", "Parameter", "The", "To", "True", "_", "__init__", "a", "act", "activation", "and", "apply_mask_to_padding_states", "a...
qwen3_5_moe/modeling_qwen3_5_moe.py:rotate_half
[ 0.00002049485374300275, 0.013860220089554787, 0.03434192016720772, 0.002974055241793394, 0.0003507140791043639, 0.028506038710474968, 0.01975221559405327, -0.02008890174329281, 0.014477476477622986, 0.018742159008979797, -0.001487027620896697, -0.01217679213732481, 0.00022445700597018003, ...
[ "Model_half", "cat", "def", "dim", "return", "shape", "torch", "x", "x1", "x2" ]
qwen3_5_moe/modeling_qwen3_5_moe.py:apply_rotary_pos_emb
[ -0.00020313466666266322, 0.028396088629961014, 0.021097494289278984, -0.006728390231728554, -0.0008695589494891465, 0.02337830513715744, 0.04196690768003464, -0.0029650533106178045, 0.011575112119317055, 0.03329982981085777, 0.0030790939927101135, 0.00803985632956028, -0.0009693444008007646,...
[ "Model_rotary_pos_emb", "cat", "cos", "def", "dim", "k", "k_embed", "k_pass", "k_rot", "q", "q_embed", "q_pass", "q_rot", "return", "rotary_dim", "rotate_half", "shape", "sin", "torch", "unsqueeze", "unsqueeze_dim" ]
qwen3_5_moe/modeling_qwen3_5_moe.py:repeat_kv
[ -0.00025096136960200965, -0.0023088448215276003, -0.004072744864970446, -0.009292741306126118, -0.000559285341296345, 0.03143470734357834, 0.009694280102849007, -0.058739304542541504, 0.011185707524418831, 0.05162634328007698, 0.005736260209232569, -0.02317449077963829, 0.0006775957299396396...
[ "Model_kv", "None", "batch", "def", "expand", "head_dim", "hidden_states", "if", "n_rep", "num_key_value_heads", "reshape", "return", "shape", "slen" ]
qwen3_5_moe/modeling_qwen3_5_moe.py:eager_attention_forward
[ 0, 0.020139193162322044, 0.013972979038953781, -0.018215786665678024, 0, 0.038468122482299805, 0.0527239553630352, -0.03326360881328583, 0.020704900845885277, 0.007297629024833441, 0.0280591007322073, 0.020704900845885277, 0.002489113714545965, -0.017876362428069115, -0.03484759107232094...
[ "Model_attention_forward", "None", "attention_mask", "attn_output", "attn_weights", "contiguous", "def", "dim", "dropout", "dtype", "float32", "functional", "if", "is", "key", "key_states", "kwargs", "matmul", "module", "nn", "not", "num_key_value_groups", "p", "query",...
qwen3_5_moe/modeling_qwen3_5_moe.py:Qwen3_5MoeAttention
[ -0.0002379890502197668, 0.022560646757483482, 0.00984880980104208, 0.004981665406376123, -0.001016374328173697, 0.027256008237600327, 0.04855692386627197, -0.01969761960208416, 0.0007121777744032443, 0.026225319132208824, 0.014429651200771332, -0.0037505642976611853, -0.00123110122513026, ...
[ "ALL_ATTENTION_FUNCTIONS", "Linear", "Model_5MoeAttention", "Model_5MoeRMSNorm", "Module", "None", "Tensor", "True", "__init__", "_attn_implementation", "apply_rotary_pos_emb", "attention_dropout", "attention_interface", "attention_mask", "attn_output", "attn_weights", "cache_kwargs"...
qwen3_5_moe/modeling_qwen3_5_moe.py:Qwen3_5MoeMLP
[ -0.0002928817120846361, 0.023604096844792366, 0.019091548398137093, 0.03216636925935745, -0.0010124306427314878, 0.05322492495179176, 0.04026581346988678, 0, -0.0038472365122288465, 0.001041357172653079, 0.017240246757864952, -0.04836525768041611, -0.0030372918117791414, 0.0220999140292406...
[ "ACT2FN", "Linear", "Model_5MoeMLP", "Module", "__init__", "act_fn", "class", "config", "def", "down_proj", "forward", "gate_proj", "hidden_act", "hidden_size", "intermediate_size", "nn", "return", "self", "super", "up_proj", "x" ]
qwen3_5_moe/modeling_qwen3_5_moe.py:Qwen3_5MoeExperts
[ -0.0003715667990036309, 0.03194017335772514, -0.005391361657530069, 0.002841393230482936, -0.0012531272368505597, 0.05735243111848831, 0.06154894828796387, -0.015503806993365288, -0.004517086781561375, -0.015037527307868004, 0.014454677700996399, -0.027160804718732834, -0.0019671183545142412...
[ "ACT2FN", "Model_5MoeExperts", "Module", "None", "Parameter", "__init__", "act_fn", "chunk", "class", "config", "continue", "current_hidden_states", "current_state", "def", "dim", "down_proj", "dtype", "empty", "expert_hit", "expert_idx", "expert_mask", "final_hidden_states...
qwen3_5_moe/modeling_qwen3_5_moe.py:Qwen3_5MoeTopKRouter
[ -0.00037748616887256503, 0.04006877541542053, 0.0032997815869748592, -0.004949672147631645, -0.0017603745218366385, 0.0551534928381443, 0.07448078691959381, -0.024748362600803375, -0.004596124403178692, 0.0210950318723917, 0.020859332755208015, -0.029462335631251335, -0.0019150518346577883, ...
[ "F", "Model_5MoeTopKRouter", "Module", "Parameter", "True", "__init__", "class", "config", "def", "dim", "dtype", "float", "forward", "functional", "hidden_dim", "hidden_size", "hidden_states", "keepdim", "linear", "nn", "num_experts", "num_experts_per_tok", "reshape", ...
qwen3_5_moe/modeling_qwen3_5_moe.py:Qwen3_5MoeSparseMoeBlock
[ -0.00045012388727627695, 0.02630806714296341, 0.001964177004992962, 0.007916228845715523, -0.0018228157423436642, 0.05118764191865921, 0.03975970298051834, -0.04142627865076065, -0.0032438680063933134, -0.008570954203605652, 0.020356016233563423, -0.03618847206234932, -0.00029202253790572286...
[ "F", "Linear", "Model_5MoeExperts", "Model_5MoeMLP", "Model_5MoeSparseMoeBlock", "Model_5MoeTopKRouter", "Module", "_", "__init__", "batch_size", "class", "config", "def", "expert_output", "experts", "forward", "gate", "hidden_dim", "hidden_size", "hidden_states", "hidden_sta...
qwen3_5_moe/modeling_qwen3_5_moe.py:Qwen3_5MoeRMSNorm
[ -0.00016307354962918907, 0.04471245035529137, 0.02121560275554657, 0.057715561240911484, -0.0007948721759021282, 0.05566243827342987, 0.024523410946130753, 0.000762792129535228, 0.006301946472376585, 0.047449950128793716, 0.010778894647955894, 0.000919627898838371, 0.0022242162376642227, 0...
[ "Model_5MoeRMSNorm", "Module", "Parameter", "True", "__init__", "_norm", "class", "def", "dim", "eps", "extra_repr", "f", "float", "forward", "keepdim", "mean", "nn", "output", "pow", "return", "rsqrt", "self", "shape", "super", "torch", "tuple", "type_as", "wei...
qwen3_5_moe/modeling_qwen3_5_moe.py:Qwen3_5MoeDecoderLayer
[ -0.0002975325915031135, 0.018239106982946396, 0.012790316715836525, 0.014625698328018188, -0.001218808232806623, 0.03991955518722534, 0.036937057971954346, -0.03510167822241783, 0.00047676911344751716, 0.014683054760098457, 0.006825326010584831, 0.002839106135070324, -0.00008379306382266805,...
[ "GradientCheckpointingLayer", "Model_5MoeAttention", "Model_5MoeDecoderLayer", "Model_5MoeGatedDeltaNet", "Model_5MoeRMSNorm", "Model_5MoeSparseMoeBlock", "None", "Tensor", "_", "__init__", "attention_mask", "cache_params", "cache_position", "class", "config", "def", "elif", "eps",...
qwen3_5_moe/modeling_qwen3_5_moe.py:Qwen3_5MoePreTrainedModel
[ -0.00030522982706315815, 0.0264293123036623, -0.01321465615183115, 0.017236508429050446, -0.0010988273425027728, 0.04481491819024086, 0.043435998260974884, -0.023786380887031555, -0.002585476031526923, 0.00884807389229536, 0.004251671954989433, -0.024705661460757256, -0.002312564756721258, ...
[ "A_log", "Model_5MoeAttention", "Model_5MoeConfig", "Model_5MoeDecoderLayer", "Model_5MoeExperts", "Model_5MoeGatedDeltaNet", "Model_5MoePreTrainedModel", "Model_5MoeRMSNorm", "Model_5MoeSparseMoeBlock", "Model_5MoeTopKRouter", "Model_5MoeVisionBlock", "Model_5MoeVisionRotaryEmbedding", "Out...
qwen3_5_moe/modeling_qwen3_5_moe.py:Qwen3_5MoeVisionMLP
[ -0.00022984776296652853, 0.02679775469005108, 0.03192923963069916, 0.030560843646526337, -0.0006841979920864105, 0.045841265469789505, 0.035806361585855484, -0.026911787688732147, -0.00022984776296652853, 0.009236672893166542, 0.03443796560168266, -0.03329763561487198, -0.0000161472507897997...
[ "ACT2FN", "Linear", "Model_5MoeVisionMLP", "Module", "__init__", "act_fn", "class", "config", "def", "forward", "hidden_act", "hidden_size", "hidden_state", "intermediate_size", "linear_fc1", "linear_fc2", "nn", "return", "self", "super" ]