{"0": "pop2piano/modeling_pop2piano.py:Pop2PianoLayerNorm", "1": "pop2piano/modeling_pop2piano.py:Pop2PianoDenseActDense", "2": "pop2piano/modeling_pop2piano.py:Pop2PianoDenseGatedActDense", "3": "pop2piano/modeling_pop2piano.py:Pop2PianoLayerFF", "4": "pop2piano/modeling_pop2piano.py:Pop2PianoAttention", "5": "pop2piano/modeling_pop2piano.py:Pop2PianoLayerSelfAttention", "6": "pop2piano/modeling_pop2piano.py:Pop2PianoLayerCrossAttention", "7": "pop2piano/modeling_pop2piano.py:Pop2PianoBlock", "8": "pop2piano/modeling_pop2piano.py:Pop2PianoPreTrainedModel", "9": "pop2piano/modeling_pop2piano.py:Pop2PianoStack", "10": "pop2piano/modeling_pop2piano.py:Pop2PianoConcatEmbeddingToMel", "11": "pop2piano/modeling_pop2piano.py:Pop2PianoForConditionalGeneration", "12": "blt/modeling_blt.py:BltMLP", "13": "blt/modeling_blt.py:BltRMSNorm", "14": "blt/modeling_blt.py:BltRotaryEmbedding", "15": "blt/modeling_blt.py:BltTransformerLayer", "16": "blt/modeling_blt.py:repeat_kv", "17": "blt/modeling_blt.py:eager_attention_forward", "18": "blt/modeling_blt.py:rotate_half", "19": "blt/modeling_blt.py:apply_rotary_pos_emb", "20": "blt/modeling_blt.py:BltSelfAttention", "21": "blt/modeling_blt.py:BltCrossAttention", "22": "blt/modeling_blt.py:BltPreTrainedModel", "23": "blt/modeling_blt.py:BltLocalEncoder", "24": "blt/modeling_blt.py:BltLocalDecoder", "25": "blt/modeling_blt.py:BltGlobalTransformer", "26": "blt/modeling_blt.py:process_patch_lengths", "27": "blt/modeling_blt.py:BltPatcher", "28": "blt/modeling_blt.py:rolling_polynomial_hash", "29": "blt/modeling_blt.py:byte_group_hash_function", "30": "blt/modeling_blt.py:compute_hash_embeddings", "31": "blt/modeling_blt.py:_prepare_patch_cross_attention_mask", "32": "blt/modeling_blt.py:BltModel", "33": "blt/modeling_blt.py:BltForCausalLM", "34": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2ForPreTrainingOutput", "35": "wav2vec2/modeling_wav2vec2.py:_compute_mask_indices", "36": "wav2vec2/modeling_wav2vec2.py:_sample_negative_indices", "37": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2NoLayerNormConvLayer", "38": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2LayerNormConvLayer", "39": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2GroupNormConvLayer", "40": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2PositionalConvEmbedding", "41": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2SamePadLayer", "42": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2FeatureEncoder", "43": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2FeatureExtractor", "44": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2FeatureProjection", "45": "wav2vec2/modeling_wav2vec2.py:eager_attention_forward", "46": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2Attention", "47": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2FeedForward", "48": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2EncoderLayer", "49": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2EncoderLayerStableLayerNorm", "50": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2Encoder", "51": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2EncoderStableLayerNorm", "52": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2GumbelVectorQuantizer", "53": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2Adapter", "54": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2AdapterLayer", "55": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2AttnAdapterLayer", "56": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2PreTrainedModel", "57": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2Model", "58": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2ForPreTraining", "59": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2ForMaskedLM", "60": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2ForCTC", "61": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2ForSequenceClassification", "62": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2ForAudioFrameClassification", "63": "wav2vec2/modeling_wav2vec2.py:AMSoftmaxLoss", "64": "wav2vec2/modeling_wav2vec2.py:TDNNLayer", "65": "wav2vec2/modeling_wav2vec2.py:Wav2Vec2ForXVector", "66": "prophetnet/modeling_prophetnet.py:softmax", "67": "prophetnet/modeling_prophetnet.py:ngram_attention_bias", "68": "prophetnet/modeling_prophetnet.py:compute_relative_buckets", "69": "prophetnet/modeling_prophetnet.py:compute_all_stream_relative_buckets", "70": "prophetnet/modeling_prophetnet.py:ProphetNetSeq2SeqLMOutput", "71": "prophetnet/modeling_prophetnet.py:ProphetNetSeq2SeqModelOutput", "72": "prophetnet/modeling_prophetnet.py:ProphetNetDecoderModelOutput", "73": "prophetnet/modeling_prophetnet.py:ProphetNetDecoderLMOutput", "74": "prophetnet/modeling_prophetnet.py:ProphetNetPreTrainedModel", "75": "prophetnet/modeling_prophetnet.py:ProphetNetPositionalEmbeddings", "76": "prophetnet/modeling_prophetnet.py:ProphetNetAttention", "77": "prophetnet/modeling_prophetnet.py:ProphetNetFeedForward", "78": "prophetnet/modeling_prophetnet.py:ProphetNetNgramSelfAttention", "79": "prophetnet/modeling_prophetnet.py:ProphetNetEncoderLayer", "80": "prophetnet/modeling_prophetnet.py:ProphetNetDecoderLayer", "81": "prophetnet/modeling_prophetnet.py:ProphetNetEncoder", "82": "prophetnet/modeling_prophetnet.py:ProphetNetDecoder", "83": "prophetnet/modeling_prophetnet.py:ProphetNetModel", "84": "prophetnet/modeling_prophetnet.py:ProphetNetForConditionalGeneration", "85": "prophetnet/modeling_prophetnet.py:ProphetNetForCausalLM", "86": "prophetnet/modeling_prophetnet.py:ProphetNetDecoderWrapper", "87": "qwen2_moe/modeling_qwen2_moe.py:load_balancing_loss_func", "88": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeRMSNorm", "89": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeRotaryEmbedding", "90": "qwen2_moe/modeling_qwen2_moe.py:rotate_half", "91": "qwen2_moe/modeling_qwen2_moe.py:apply_rotary_pos_emb", "92": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeMLP", "93": "qwen2_moe/modeling_qwen2_moe.py:repeat_kv", "94": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeAttention", "95": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeFlashAttention2", "96": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeSdpaAttention", "97": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeSparseMoeBlock", "98": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeDecoderLayer", "99": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoePreTrainedModel", "100": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeModel", "101": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeForCausalLM", "102": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeForSequenceClassification", "103": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeForTokenClassification", "104": "qwen2_moe/modeling_qwen2_moe.py:Qwen2MoeForQuestionAnswering", "105": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackbonePatchEmbeddings", "106": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneEmbeddings", "107": "vitpose_backbone/modeling_vitpose_backbone.py:eager_attention_forward", "108": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneSelfAttention", "109": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneSelfOutput", "110": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneAttention", "111": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneMoeMLP", "112": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneMLP", "113": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneLayer", "114": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackboneEncoder", "115": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackbonePreTrainedModel", "116": "vitpose_backbone/modeling_vitpose_backbone.py:VitPoseBackbone", "117": "sam2_video/modeling_sam2_video.py:Sam2VideoInferenceCache", "118": "sam2_video/modeling_sam2_video.py:Sam2VideoInferenceSession", "119": "sam2_video/modeling_sam2_video.py:Sam2VideoLayerNorm", "120": "sam2_video/modeling_sam2_video.py:Sam2VideoPositionEmbeddingSine", "121": "sam2_video/modeling_sam2_video.py:eager_attention_forward", "122": "sam2_video/modeling_sam2_video.py:Sam2VideoAttention", "123": "sam2_video/modeling_sam2_video.py:Sam2VideoTwoWayAttentionBlock", "124": "sam2_video/modeling_sam2_video.py:Sam2VideoFeedForward", "125": "sam2_video/modeling_sam2_video.py:Sam2VideoImageSegmentationOutput", "126": "sam2_video/modeling_sam2_video.py:Sam2VideoSegmentationOutput", "127": "sam2_video/modeling_sam2_video.py:Sam2VideoPreTrainedModel", "128": "sam2_video/modeling_sam2_video.py:Sam2VideoVisionRotaryEmbedding", "129": "sam2_video/modeling_sam2_video.py:rotate_pairwise", "130": "sam2_video/modeling_sam2_video.py:apply_rotary_pos_emb_2d", "131": "sam2_video/modeling_sam2_video.py:Sam2VideoRoPEAttention", "132": "sam2_video/modeling_sam2_video.py:Sam2VideoMemoryAttentionLayer", "133": "sam2_video/modeling_sam2_video.py:Sam2VideoMemoryAttention", "134": "sam2_video/modeling_sam2_video.py:Sam2VideoMemoryFuserCXBlock", "135": "sam2_video/modeling_sam2_video.py:Sam2VideoMemoryFuser", "136": "sam2_video/modeling_sam2_video.py:Sam2VideoMaskDownSamplerLayer", "137": "sam2_video/modeling_sam2_video.py:Sam2VideoMaskDownSampler", "138": "sam2_video/modeling_sam2_video.py:Sam2VideoMemoryEncoder", "139": "sam2_video/modeling_sam2_video.py:Sam2VideoVisionEncoderOutput", "140": "sam2_video/modeling_sam2_video.py:Sam2VideoPositionalEmbedding", "141": "sam2_video/modeling_sam2_video.py:Sam2VideoMaskEmbedding", "142": "sam2_video/modeling_sam2_video.py:Sam2VideoPromptEncoder", "143": "sam2_video/modeling_sam2_video.py:Sam2VideoTwoWayTransformer", "144": "sam2_video/modeling_sam2_video.py:Sam2VideoMaskDecoder", "145": "sam2_video/modeling_sam2_video.py:get_1d_sine_pe", "146": "sam2_video/modeling_sam2_video.py:Sam2VideoModel", "147": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerGatedAttention", "148": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerBatchNorm", "149": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerPositionalEncoding", "150": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerNormLayer", "151": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerMLP", "152": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerChannelFeatureMixerBlock", "153": "patchtsmixer/modeling_patchtsmixer.py:eager_attention_forward", "154": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerAttention", "155": "patchtsmixer/modeling_patchtsmixer.py:PatchMixerBlock", "156": "patchtsmixer/modeling_patchtsmixer.py:FeatureMixerBlock", "157": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerLayer", "158": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerBlock", "159": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForPredictionHead", "160": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerLinearHead", "161": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerPreTrainedModel", "162": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerPretrainHead", "163": "patchtsmixer/modeling_patchtsmixer.py:random_masking", "164": "patchtsmixer/modeling_patchtsmixer.py:forecast_masking", "165": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerPatchify", "166": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerMasking", "167": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerStdScaler", "168": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerMeanScaler", "169": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerNOPScaler", "170": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerEncoderOutput", "171": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerEncoder", "172": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerModelOutput", "173": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerModel", "174": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForPreTrainingOutput", "175": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForPretraining", "176": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForPredictionOutput", "177": "patchtsmixer/modeling_patchtsmixer.py:SamplePatchTSMixerPredictionOutput", "178": "patchtsmixer/modeling_patchtsmixer.py:SamplePatchTSMixerRegressionOutput", "179": "patchtsmixer/modeling_patchtsmixer.py:nll", "180": "patchtsmixer/modeling_patchtsmixer.py:weighted_average", "181": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForPrediction", "182": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForTimeSeriesClassificationOutput", "183": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForTimeSeriesClassification", "184": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForRegressionOutput", "185": "patchtsmixer/modeling_patchtsmixer.py:InjectScalerStatistics4D", "186": "patchtsmixer/modeling_patchtsmixer.py:PatchTSMixerForRegression", "187": "doge/modeling_doge.py:DogeRMSNorm", "188": "doge/modeling_doge.py:DogeRotaryEmbedding", "189": "doge/modeling_doge.py:rotate_half", "190": "doge/modeling_doge.py:apply_rotary_pos_emb", "191": "doge/modeling_doge.py:repeat_kv", "192": "doge/modeling_doge.py:eager_attention_forward", "193": "doge/modeling_doge.py:flex_attention_forward", "194": "doge/modeling_doge.py:DogeAttention", "195": "doge/modeling_doge.py:DogeMLP", "196": "doge/modeling_doge.py:DogeCDMoE", "197": "doge/modeling_doge.py:DogeDecoderLayer", "198": "doge/modeling_doge.py:DogePreTrainedModel", "199": "doge/modeling_doge.py:DogeModel", "200": "doge/modeling_doge.py:load_balancing_loss_func", "201": "doge/modeling_doge.py:DogeForCausalLM", "202": "doge/modeling_doge.py:DogeForSequenceClassification", "203": "dac/modeling_dac.py:DacOutput", "204": "dac/modeling_dac.py:DacEncoderOutput", "205": "dac/modeling_dac.py:DacDecoderOutput", "206": "dac/modeling_dac.py:Snake1d", "207": "dac/modeling_dac.py:DacVectorQuantize", "208": "dac/modeling_dac.py:DacResidualUnit", "209": "dac/modeling_dac.py:DacEncoderBlock", "210": "dac/modeling_dac.py:DacDecoderBlock", "211": "dac/modeling_dac.py:DacResidualVectorQuantize", "212": "dac/modeling_dac.py:DacDecoder", "213": "dac/modeling_dac.py:DacEncoder", "214": "dac/modeling_dac.py:DacPreTrainedModel", "215": "dac/modeling_dac.py:DacModel", "216": "chinese_clip/modeling_chinese_clip.py:contrastive_loss", "217": "chinese_clip/modeling_chinese_clip.py:chinese_clip_loss", "218": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPOutput", "219": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextEmbeddings", "220": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPVisionEmbeddings", "221": "chinese_clip/modeling_chinese_clip.py:eager_attention_forward", "222": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextSelfAttention", "223": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextSelfOutput", "224": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextAttention", "225": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPVisionAttention", "226": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextIntermediate", "227": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextOutput", "228": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPVisionMLP", "229": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextLayer", "230": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPVisionLayer", "231": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextPooler", "232": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPPreTrainedModel", "233": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextEncoder", "234": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPVisionEncoder", "235": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPVisionTransformer", "236": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPTextModel", "237": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPVisionModel", "238": "chinese_clip/modeling_chinese_clip.py:ChineseCLIPModel", "239": "convbert/modeling_convbert.py:ConvBertEmbeddings", "240": "convbert/modeling_convbert.py:ConvBertPreTrainedModel", "241": "convbert/modeling_convbert.py:SeparableConv1D", "242": "convbert/modeling_convbert.py:ConvBertSelfAttention", "243": "convbert/modeling_convbert.py:ConvBertSelfOutput", "244": "convbert/modeling_convbert.py:ConvBertAttention", "245": "convbert/modeling_convbert.py:GroupedLinearLayer", "246": "convbert/modeling_convbert.py:ConvBertIntermediate", "247": "convbert/modeling_convbert.py:ConvBertOutput", "248": "convbert/modeling_convbert.py:ConvBertLayer", "249": "convbert/modeling_convbert.py:ConvBertEncoder", "250": "convbert/modeling_convbert.py:ConvBertPredictionHeadTransform", "251": "convbert/modeling_convbert.py:ConvBertSequenceSummary", "252": "convbert/modeling_convbert.py:ConvBertModel", "253": "convbert/modeling_convbert.py:ConvBertGeneratorPredictions", "254": "convbert/modeling_convbert.py:ConvBertForMaskedLM", "255": "convbert/modeling_convbert.py:ConvBertClassificationHead", "256": "convbert/modeling_convbert.py:ConvBertForSequenceClassification", "257": "convbert/modeling_convbert.py:ConvBertForMultipleChoice", "258": "convbert/modeling_convbert.py:ConvBertForTokenClassification", "259": "convbert/modeling_convbert.py:ConvBertForQuestionAnswering", "260": "xlnet/modeling_xlnet.py:XLNetRelativeAttention", "261": "xlnet/modeling_xlnet.py:XLNetFeedForward", "262": "xlnet/modeling_xlnet.py:XLNetLayer", "263": "xlnet/modeling_xlnet.py:XLNetPoolerStartLogits", "264": "xlnet/modeling_xlnet.py:XLNetPoolerEndLogits", "265": "xlnet/modeling_xlnet.py:XLNetPoolerAnswerClass", "266": "xlnet/modeling_xlnet.py:XLNetSequenceSummary", "267": "xlnet/modeling_xlnet.py:XLNetPreTrainedModel", "268": "xlnet/modeling_xlnet.py:XLNetModelOutput", "269": "xlnet/modeling_xlnet.py:XLNetLMHeadModelOutput", "270": "xlnet/modeling_xlnet.py:XLNetForSequenceClassificationOutput", "271": "xlnet/modeling_xlnet.py:XLNetForTokenClassificationOutput", "272": "xlnet/modeling_xlnet.py:XLNetForMultipleChoiceOutput", "273": "xlnet/modeling_xlnet.py:XLNetForQuestionAnsweringSimpleOutput", "274": "xlnet/modeling_xlnet.py:XLNetForQuestionAnsweringOutput", "275": "xlnet/modeling_xlnet.py:XLNetModel", "276": "xlnet/modeling_xlnet.py:XLNetLMHeadModel", "277": "xlnet/modeling_xlnet.py:XLNetForSequenceClassification", "278": "xlnet/modeling_xlnet.py:XLNetForTokenClassification", "279": "xlnet/modeling_xlnet.py:XLNetForMultipleChoice", "280": "xlnet/modeling_xlnet.py:XLNetForQuestionAnsweringSimple", "281": "xlnet/modeling_xlnet.py:XLNetForQuestionAnswering", "282": "upernet/modeling_upernet.py:UperNetConvModule", "283": "upernet/modeling_upernet.py:UperNetPyramidPoolingBlock", "284": "upernet/modeling_upernet.py:UperNetPyramidPoolingModule", "285": "upernet/modeling_upernet.py:UperNetHead", "286": "upernet/modeling_upernet.py:UperNetFCNHead", "287": "upernet/modeling_upernet.py:UperNetPreTrainedModel", "288": "upernet/modeling_upernet.py:UperNetForSemanticSegmentation", "289": "minimax/modeling_minimax.py:MiniMaxRMSNorm", "290": "minimax/modeling_minimax.py:MiniMaxCache", "291": "minimax/modeling_minimax.py:MiniMaxLightningAttention", "292": "minimax/modeling_minimax.py:rotate_half", "293": "minimax/modeling_minimax.py:apply_rotary_pos_emb", "294": "minimax/modeling_minimax.py:repeat_kv", "295": "minimax/modeling_minimax.py:eager_attention_forward", "296": "minimax/modeling_minimax.py:MiniMaxAttention", "297": "minimax/modeling_minimax.py:MiniMaxBlockSparseTop2MLP", "298": "minimax/modeling_minimax.py:MiniMaxSparseMoeBlock", "299": "minimax/modeling_minimax.py:MiniMaxDecoderLayer", "300": "minimax/modeling_minimax.py:MiniMaxPreTrainedModel", "301": "minimax/modeling_minimax.py:MiniMaxRotaryEmbedding", "302": "minimax/modeling_minimax.py:MiniMaxModel", "303": "minimax/modeling_minimax.py:load_balancing_loss_func", "304": "minimax/modeling_minimax.py:MiniMaxForCausalLM", "305": "minimax/modeling_minimax.py:MiniMaxForSequenceClassification", "306": "minimax/modeling_minimax.py:MiniMaxForTokenClassification", "307": "minimax/modeling_minimax.py:MiniMaxForQuestionAnswering", "308": "xlstm/modeling_xlstm.py:small_init_method", "309": "xlstm/modeling_xlstm.py:wang_init_method", "310": "xlstm/modeling_xlstm.py:xLSTMPreTrainedModel", "311": "xlstm/modeling_xlstm.py:xLSTMCache", "312": "xlstm/modeling_xlstm.py:xLSTMOutput", "313": "xlstm/modeling_xlstm.py:xLSTMModel", "314": "xlstm/modeling_xlstm.py:xLSTMCausalLMOutput", "315": "xlstm/modeling_xlstm.py:xLSTMForCausalLM", "316": "seed_oss/modeling_seed_oss.py:SeedOssRMSNorm", "317": "seed_oss/modeling_seed_oss.py:SeedOssMLP", "318": "seed_oss/modeling_seed_oss.py:rotate_half", "319": "seed_oss/modeling_seed_oss.py:apply_rotary_pos_emb", "320": "seed_oss/modeling_seed_oss.py:repeat_kv", "321": "seed_oss/modeling_seed_oss.py:eager_attention_forward", "322": "seed_oss/modeling_seed_oss.py:SeedOssAttention", "323": "seed_oss/modeling_seed_oss.py:SeedOssDecoderLayer", "324": "seed_oss/modeling_seed_oss.py:SeedOssPreTrainedModel", "325": "seed_oss/modeling_seed_oss.py:SeedOssRotaryEmbedding", "326": "seed_oss/modeling_seed_oss.py:SeedOssModel", "327": "seed_oss/modeling_seed_oss.py:SeedOssForCausalLM", "328": "seed_oss/modeling_seed_oss.py:SeedOssForSequenceClassification", "329": "seed_oss/modeling_seed_oss.py:SeedOssForTokenClassification", "330": "seed_oss/modeling_seed_oss.py:SeedOssForQuestionAnswering", "331": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerModelOutput", "332": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerWithHifiGanOutput", "333": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:length_regulator", "334": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerDurationPredictor", "335": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerBatchNormConvLayer", "336": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerSpeechDecoderPostnet", "337": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerPredictorLayer", "338": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerVariancePredictor", "339": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerVarianceEmbedding", "340": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerAttention", "341": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerConvolutionModule", "342": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerEncoderLayer", "343": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerMultiLayeredConv1d", "344": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerRelPositionalEncoding", "345": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerEncoder", "346": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerLoss", "347": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerPreTrainedModel", "348": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerModel", "349": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:HifiGanResidualBlock", "350": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerHifiGan", "351": "fastspeech2_conformer/modeling_fastspeech2_conformer.py:FastSpeech2ConformerWithHifiGan", "352": "bert/modeling_bert.py:BertEmbeddings", "353": "bert/modeling_bert.py:eager_attention_forward", "354": "bert/modeling_bert.py:BertSelfAttention", "355": "bert/modeling_bert.py:BertCrossAttention", "356": "bert/modeling_bert.py:BertSelfOutput", "357": "bert/modeling_bert.py:BertAttention", "358": "bert/modeling_bert.py:BertIntermediate", "359": "bert/modeling_bert.py:BertOutput", "360": "bert/modeling_bert.py:BertLayer", "361": "bert/modeling_bert.py:BertEncoder", "362": "bert/modeling_bert.py:BertPooler", "363": "bert/modeling_bert.py:BertPredictionHeadTransform", "364": "bert/modeling_bert.py:BertLMPredictionHead", "365": "bert/modeling_bert.py:BertOnlyMLMHead", "366": "bert/modeling_bert.py:BertOnlyNSPHead", "367": "bert/modeling_bert.py:BertPreTrainingHeads", "368": "bert/modeling_bert.py:BertPreTrainedModel", "369": "bert/modeling_bert.py:BertForPreTrainingOutput", "370": "bert/modeling_bert.py:BertModel", "371": "bert/modeling_bert.py:BertForPreTraining", "372": "bert/modeling_bert.py:BertLMHeadModel", "373": "bert/modeling_bert.py:BertForMaskedLM", "374": "bert/modeling_bert.py:BertForNextSentencePrediction", "375": "bert/modeling_bert.py:BertForSequenceClassification", "376": "bert/modeling_bert.py:BertForMultipleChoice", "377": "bert/modeling_bert.py:BertForTokenClassification", "378": "bert/modeling_bert.py:BertForQuestionAnswering", "379": "stablelm/modeling_stablelm.py:StableLmRotaryEmbedding", "380": "stablelm/modeling_stablelm.py:rotate_half", "381": "stablelm/modeling_stablelm.py:apply_rotary_pos_emb", "382": "stablelm/modeling_stablelm.py:StableLmMLP", "383": "stablelm/modeling_stablelm.py:StableLmLayerNormPerHead", "384": "stablelm/modeling_stablelm.py:repeat_kv", "385": "stablelm/modeling_stablelm.py:StableLmAttention", "386": "stablelm/modeling_stablelm.py:StableLmSdpaAttention", "387": "stablelm/modeling_stablelm.py:StableLmFlashAttention2", "388": "stablelm/modeling_stablelm.py:StableLmDecoderLayer", "389": "stablelm/modeling_stablelm.py:StableLmPreTrainedModel", "390": "stablelm/modeling_stablelm.py:StableLmModel", "391": "stablelm/modeling_stablelm.py:StableLmForCausalLM", "392": "stablelm/modeling_stablelm.py:StableLmForSequenceClassification", "393": "stablelm/modeling_stablelm.py:StableLmForTokenClassification", "394": "llava/modeling_llava.py:LlavaModelOutputWithPast", "395": "llava/modeling_llava.py:LlavaCausalLMOutputWithPast", "396": "llava/modeling_llava.py:LlavaMultiModalProjector", "397": "llava/modeling_llava.py:LlavaPreTrainedModel", "398": "llava/modeling_llava.py:LlavaModel", "399": "llava/modeling_llava.py:LlavaForConditionalGeneration", "400": "roformer/modeling_roformer.py:RoFormerSinusoidalPositionalEmbedding", "401": "roformer/modeling_roformer.py:RoFormerEmbeddings", "402": "roformer/modeling_roformer.py:RoFormerSelfAttention", "403": "roformer/modeling_roformer.py:RoFormerSelfOutput", "404": "roformer/modeling_roformer.py:RoFormerAttention", "405": "roformer/modeling_roformer.py:RoFormerIntermediate", "406": "roformer/modeling_roformer.py:RoFormerOutput", "407": "roformer/modeling_roformer.py:RoFormerLayer", "408": "roformer/modeling_roformer.py:RoFormerEncoder", "409": "roformer/modeling_roformer.py:RoFormerSequenceSummary", "410": "roformer/modeling_roformer.py:RoFormerPredictionHeadTransform", "411": "roformer/modeling_roformer.py:RoFormerLMPredictionHead", "412": "roformer/modeling_roformer.py:RoFormerOnlyMLMHead", "413": "roformer/modeling_roformer.py:RoFormerPreTrainedModel", "414": "roformer/modeling_roformer.py:RoFormerModel", "415": "roformer/modeling_roformer.py:RoFormerForMaskedLM", "416": "roformer/modeling_roformer.py:RoFormerForCausalLM", "417": "roformer/modeling_roformer.py:RoFormerClassificationHead", "418": "roformer/modeling_roformer.py:RoFormerForSequenceClassification", "419": "roformer/modeling_roformer.py:RoFormerForMultipleChoice", "420": "roformer/modeling_roformer.py:RoFormerForTokenClassification", "421": "roformer/modeling_roformer.py:RoFormerForQuestionAnswering", "422": "gpt_neo/modeling_gpt_neo.py:GPTNeoSelfAttention", "423": "gpt_neo/modeling_gpt_neo.py:GPTNeoFlashAttention2", "424": "gpt_neo/modeling_gpt_neo.py:GPTNeoAttention", "425": "gpt_neo/modeling_gpt_neo.py:GPTNeoMLP", "426": "gpt_neo/modeling_gpt_neo.py:GPTNeoBlock", "427": "gpt_neo/modeling_gpt_neo.py:GPTNeoPreTrainedModel", "428": "gpt_neo/modeling_gpt_neo.py:GPTNeoModel", "429": "gpt_neo/modeling_gpt_neo.py:GPTNeoForCausalLM", "430": "gpt_neo/modeling_gpt_neo.py:GPTNeoForSequenceClassification", "431": "gpt_neo/modeling_gpt_neo.py:GPTNeoForTokenClassification", "432": "gpt_neo/modeling_gpt_neo.py:GPTNeoForQuestionAnswering", "433": "phi/modeling_phi.py:rotate_half", "434": "phi/modeling_phi.py:apply_rotary_pos_emb", "435": "phi/modeling_phi.py:repeat_kv", "436": "phi/modeling_phi.py:eager_attention_forward", "437": "phi/modeling_phi.py:PhiAttention", "438": "phi/modeling_phi.py:PhiMLP", "439": "phi/modeling_phi.py:PhiDecoderLayer", "440": "phi/modeling_phi.py:PhiRotaryEmbedding", "441": "phi/modeling_phi.py:PhiPreTrainedModel", "442": "phi/modeling_phi.py:PhiModel", "443": "phi/modeling_phi.py:PhiForCausalLM", "444": "phi/modeling_phi.py:PhiForSequenceClassification", "445": "phi/modeling_phi.py:PhiForTokenClassification", "446": "vit_msn/modeling_vit_msn.py:ViTMSNEmbeddings", "447": "vit_msn/modeling_vit_msn.py:ViTMSNPatchEmbeddings", "448": "vit_msn/modeling_vit_msn.py:eager_attention_forward", "449": "vit_msn/modeling_vit_msn.py:ViTMSNSelfAttention", "450": "vit_msn/modeling_vit_msn.py:ViTMSNSelfOutput", "451": "vit_msn/modeling_vit_msn.py:ViTMSNAttention", "452": "vit_msn/modeling_vit_msn.py:ViTMSNIntermediate", "453": "vit_msn/modeling_vit_msn.py:ViTMSNOutput", "454": "vit_msn/modeling_vit_msn.py:ViTMSNLayer", "455": "vit_msn/modeling_vit_msn.py:ViTMSNEncoder", "456": "vit_msn/modeling_vit_msn.py:ViTMSNPreTrainedModel", "457": "vit_msn/modeling_vit_msn.py:ViTMSNModel", "458": "vit_msn/modeling_vit_msn.py:ViTMSNForImageClassification", "459": "xglm/modeling_xglm.py:XGLMScaledWordEmbedding", "460": "xglm/modeling_xglm.py:XGLMSinusoidalPositionalEmbedding", "461": "xglm/modeling_xglm.py:XGLMAttention", "462": "xglm/modeling_xglm.py:XGLMDecoderLayer", "463": "xglm/modeling_xglm.py:XGLMPreTrainedModel", "464": "xglm/modeling_xglm.py:XGLMModel", "465": "xglm/modeling_xglm.py:XGLMForCausalLM", "466": "swin2sr/modeling_swin2sr.py:Swin2SREncoderOutput", "467": "swin2sr/modeling_swin2sr.py:window_partition", "468": "swin2sr/modeling_swin2sr.py:window_reverse", "469": "swin2sr/modeling_swin2sr.py:drop_path", "470": "swin2sr/modeling_swin2sr.py:Swin2SRDropPath", "471": "swin2sr/modeling_swin2sr.py:Swin2SREmbeddings", "472": "swin2sr/modeling_swin2sr.py:Swin2SRPatchEmbeddings", "473": "swin2sr/modeling_swin2sr.py:Swin2SRPatchUnEmbeddings", "474": "swin2sr/modeling_swin2sr.py:Swin2SRPatchMerging", "475": "swin2sr/modeling_swin2sr.py:Swin2SRSelfAttention", "476": "swin2sr/modeling_swin2sr.py:Swin2SRSelfOutput", "477": "swin2sr/modeling_swin2sr.py:Swin2SRAttention", "478": "swin2sr/modeling_swin2sr.py:Swin2SRIntermediate", "479": "swin2sr/modeling_swin2sr.py:Swin2SROutput", "480": "swin2sr/modeling_swin2sr.py:Swin2SRLayer", "481": "swin2sr/modeling_swin2sr.py:Swin2SRStage", "482": "swin2sr/modeling_swin2sr.py:Swin2SREncoder", "483": "swin2sr/modeling_swin2sr.py:Swin2SRPreTrainedModel", "484": "swin2sr/modeling_swin2sr.py:Swin2SRModel", "485": "swin2sr/modeling_swin2sr.py:Upsample", "486": "swin2sr/modeling_swin2sr.py:UpsampleOneStep", "487": "swin2sr/modeling_swin2sr.py:PixelShuffleUpsampler", "488": "swin2sr/modeling_swin2sr.py:NearestConvUpsampler", "489": "swin2sr/modeling_swin2sr.py:PixelShuffleAuxUpsampler", "490": "swin2sr/modeling_swin2sr.py:Swin2SRForImageSuperResolution", "491": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLMLP", "492": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VisionPatchEmbed", "493": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VisionRotaryEmbedding", "494": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLPatchMerger", "495": "qwen2_5_vl/modeling_qwen2_5_vl.py:rotate_half", "496": "qwen2_5_vl/modeling_qwen2_5_vl.py:apply_rotary_pos_emb_vision", "497": "qwen2_5_vl/modeling_qwen2_5_vl.py:repeat_kv", "498": "qwen2_5_vl/modeling_qwen2_5_vl.py:eager_attention_forward", "499": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLVisionAttention", "500": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLVisionBlock", "501": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLPreTrainedModel", "502": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VisionTransformerPretrainedModel", "503": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLModelOutputWithPast", "504": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLRotaryEmbedding", "505": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2MLP", "506": "qwen2_5_vl/modeling_qwen2_5_vl.py:apply_multimodal_rotary_pos_emb", "507": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLAttention", "508": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLDecoderLayer", "509": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLTextModel", "510": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLModel", "511": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLCausalLMOutputWithPast", "512": "qwen2_5_vl/modeling_qwen2_5_vl.py:Qwen2_5_VLForConditionalGeneration", "513": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeRMSNorm", "514": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeMLP", "515": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeRotaryEmbedding", "516": "ernie4_5_moe/modeling_ernie4_5_moe.py:rotate_half", "517": "ernie4_5_moe/modeling_ernie4_5_moe.py:apply_rotary_pos_emb", "518": "ernie4_5_moe/modeling_ernie4_5_moe.py:repeat_kv", "519": "ernie4_5_moe/modeling_ernie4_5_moe.py:eager_attention_forward", "520": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeAttention", "521": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeStatics", "522": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeSparseMoeBlock", "523": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeDecoderLayer", "524": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoePreTrainedModel", "525": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeModel", "526": "ernie4_5_moe/modeling_ernie4_5_moe.py:load_balancing_loss_func", "527": "ernie4_5_moe/modeling_ernie4_5_moe.py:Ernie4_5_MoeForCausalLM", "528": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoContrastiveEmbedding", "529": "mm_grounding_dino/modeling_mm_grounding_dino.py:MultiScaleDeformableAttention", "530": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoLearnedPositionEmbedding", "531": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoMultiscaleDeformableAttention", "532": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoBiMultiHeadAttention", "533": "mm_grounding_dino/modeling_mm_grounding_dino.py:drop_path", "534": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoDropPath", "535": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoFusionLayer", "536": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoPreTrainedModel", "537": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoFrozenBatchNorm2d", "538": "mm_grounding_dino/modeling_mm_grounding_dino.py:replace_batch_norm", "539": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoConvEncoder", "540": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoConvModel", "541": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoEncoderOutput", "542": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoMultiheadAttention", "543": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoTextEnhancerLayer", "544": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoDeformableLayer", "545": "mm_grounding_dino/modeling_mm_grounding_dino.py:get_sine_pos_embed", "546": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoEncoderLayer", "547": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoEncoder", "548": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoDecoderOutput", "549": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoDecoderLayer", "550": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoDecoder", "551": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoModelOutput", "552": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoSinePositionEmbedding", "553": "mm_grounding_dino/modeling_mm_grounding_dino.py:build_position_encoding", "554": "mm_grounding_dino/modeling_mm_grounding_dino.py:generate_masks_with_special_tokens_and_transfer_map", "555": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoModel", "556": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoMLPPredictionHead", "557": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoObjectDetectionOutput", "558": "mm_grounding_dino/modeling_mm_grounding_dino.py:build_label_maps", "559": "mm_grounding_dino/modeling_mm_grounding_dino.py:build_text_mask", "560": "mm_grounding_dino/modeling_mm_grounding_dino.py:MMGroundingDinoForObjectDetection", "561": "umt5/modeling_umt5.py:UMT5LayerNorm", "562": "umt5/modeling_umt5.py:UMT5DenseActDense", "563": "umt5/modeling_umt5.py:UMT5DenseGatedActDense", "564": "umt5/modeling_umt5.py:UMT5LayerFF", "565": "umt5/modeling_umt5.py:UMT5Attention", "566": "umt5/modeling_umt5.py:UMT5LayerSelfAttention", "567": "umt5/modeling_umt5.py:UMT5LayerCrossAttention", "568": "umt5/modeling_umt5.py:UMT5Block", "569": "umt5/modeling_umt5.py:UMT5ClassificationHead", "570": "umt5/modeling_umt5.py:UMT5PreTrainedModel", "571": "umt5/modeling_umt5.py:UMT5Stack", "572": "umt5/modeling_umt5.py:UMT5Model", "573": "umt5/modeling_umt5.py:UMT5ForConditionalGeneration", "574": "umt5/modeling_umt5.py:UMT5EncoderModel", "575": "umt5/modeling_umt5.py:UMT5ForSequenceClassification", "576": "umt5/modeling_umt5.py:UMT5ForTokenClassification", "577": "umt5/modeling_umt5.py:UMT5ForQuestionAnswering", "578": "funnel/modeling_funnel.py:FunnelEmbeddings", "579": "funnel/modeling_funnel.py:FunnelAttentionStructure", "580": "funnel/modeling_funnel.py:_relative_shift_gather", "581": "funnel/modeling_funnel.py:FunnelRelMultiheadAttention", "582": "funnel/modeling_funnel.py:FunnelPositionwiseFFN", "583": "funnel/modeling_funnel.py:FunnelLayer", "584": "funnel/modeling_funnel.py:FunnelEncoder", "585": "funnel/modeling_funnel.py:upsample", "586": "funnel/modeling_funnel.py:FunnelDecoder", "587": "funnel/modeling_funnel.py:FunnelDiscriminatorPredictions", "588": "funnel/modeling_funnel.py:FunnelPreTrainedModel", "589": "funnel/modeling_funnel.py:FunnelClassificationHead", "590": "funnel/modeling_funnel.py:FunnelForPreTrainingOutput", "591": "funnel/modeling_funnel.py:FunnelBaseModel", "592": "funnel/modeling_funnel.py:FunnelModel", "593": "funnel/modeling_funnel.py:FunnelForPreTraining", "594": "funnel/modeling_funnel.py:FunnelForMaskedLM", "595": "funnel/modeling_funnel.py:FunnelForSequenceClassification", "596": "funnel/modeling_funnel.py:FunnelForMultipleChoice", "597": "funnel/modeling_funnel.py:FunnelForTokenClassification", "598": "funnel/modeling_funnel.py:FunnelForQuestionAnswering", "599": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3PatchEmbeddings", "600": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3TextEmbeddings", "601": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3PreTrainedModel", "602": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3SelfAttention", "603": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3SelfOutput", "604": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3Attention", "605": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3Layer", "606": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3Encoder", "607": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3Intermediate", "608": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3Output", "609": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3Model", "610": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3ClassificationHead", "611": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3ForTokenClassification", "612": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3ForQuestionAnswering", "613": "layoutlmv3/modeling_layoutlmv3.py:LayoutLMv3ForSequenceClassification", "614": "paligemma/modeling_paligemma.py:PaligemmaModelOutputWithPast", "615": "paligemma/modeling_paligemma.py:PaliGemmaCausalLMOutputWithPast", "616": "paligemma/modeling_paligemma.py:PaliGemmaMultiModalProjector", "617": "paligemma/modeling_paligemma.py:token_type_ids_mask_function", "618": "paligemma/modeling_paligemma.py:create_causal_mask_mapping", "619": "paligemma/modeling_paligemma.py:PaliGemmaPreTrainedModel", "620": "paligemma/modeling_paligemma.py:PaliGemmaModel", "621": "paligemma/modeling_paligemma.py:PaliGemmaForConditionalGeneration", "622": "nystromformer/modeling_nystromformer.py:NystromformerEmbeddings", "623": "nystromformer/modeling_nystromformer.py:NystromformerSelfAttention", "624": "nystromformer/modeling_nystromformer.py:NystromformerSelfOutput", "625": "nystromformer/modeling_nystromformer.py:NystromformerAttention", "626": "nystromformer/modeling_nystromformer.py:NystromformerIntermediate", "627": "nystromformer/modeling_nystromformer.py:NystromformerOutput", "628": "nystromformer/modeling_nystromformer.py:NystromformerLayer", "629": "nystromformer/modeling_nystromformer.py:NystromformerEncoder", "630": "nystromformer/modeling_nystromformer.py:NystromformerPredictionHeadTransform", "631": "nystromformer/modeling_nystromformer.py:NystromformerLMPredictionHead", "632": "nystromformer/modeling_nystromformer.py:NystromformerOnlyMLMHead", "633": "nystromformer/modeling_nystromformer.py:NystromformerPreTrainedModel", "634": "nystromformer/modeling_nystromformer.py:NystromformerModel", "635": "nystromformer/modeling_nystromformer.py:NystromformerForMaskedLM", "636": "nystromformer/modeling_nystromformer.py:NystromformerClassificationHead", "637": "nystromformer/modeling_nystromformer.py:NystromformerForSequenceClassification", "638": "nystromformer/modeling_nystromformer.py:NystromformerForMultipleChoice", "639": "nystromformer/modeling_nystromformer.py:NystromformerForTokenClassification", "640": "nystromformer/modeling_nystromformer.py:NystromformerForQuestionAnswering", "641": "dinov2/modeling_dinov2.py:Dinov2Embeddings", "642": "dinov2/modeling_dinov2.py:Dinov2PatchEmbeddings", "643": "dinov2/modeling_dinov2.py:eager_attention_forward", "644": "dinov2/modeling_dinov2.py:Dinov2SelfAttention", "645": "dinov2/modeling_dinov2.py:Dinov2SelfOutput", "646": "dinov2/modeling_dinov2.py:Dinov2Attention", "647": "dinov2/modeling_dinov2.py:Dinov2LayerScale", "648": "dinov2/modeling_dinov2.py:drop_path", "649": "dinov2/modeling_dinov2.py:Dinov2DropPath", "650": "dinov2/modeling_dinov2.py:Dinov2MLP", "651": "dinov2/modeling_dinov2.py:Dinov2SwiGLUFFN", "652": "dinov2/modeling_dinov2.py:Dinov2Layer", "653": "dinov2/modeling_dinov2.py:Dinov2Encoder", "654": "dinov2/modeling_dinov2.py:Dinov2PreTrainedModel", "655": "dinov2/modeling_dinov2.py:Dinov2Model", "656": "dinov2/modeling_dinov2.py:Dinov2ForImageClassification", "657": "dinov2/modeling_dinov2.py:Dinov2Backbone", "658": "lxmert/modeling_lxmert.py:GeLU", "659": "lxmert/modeling_lxmert.py:LxmertModelOutput", "660": "lxmert/modeling_lxmert.py:LxmertForQuestionAnsweringOutput", "661": "lxmert/modeling_lxmert.py:LxmertForPreTrainingOutput", "662": "lxmert/modeling_lxmert.py:LxmertEmbeddings", "663": "lxmert/modeling_lxmert.py:LxmertAttention", "664": "lxmert/modeling_lxmert.py:LxmertAttentionOutput", "665": "lxmert/modeling_lxmert.py:LxmertCrossAttentionLayer", "666": "lxmert/modeling_lxmert.py:LxmertSelfAttentionLayer", "667": "lxmert/modeling_lxmert.py:LxmertIntermediate", "668": "lxmert/modeling_lxmert.py:LxmertOutput", "669": "lxmert/modeling_lxmert.py:LxmertLayer", "670": "lxmert/modeling_lxmert.py:LxmertXLayer", "671": "lxmert/modeling_lxmert.py:LxmertVisualFeatureEncoder", "672": "lxmert/modeling_lxmert.py:LxmertEncoder", "673": "lxmert/modeling_lxmert.py:LxmertPooler", "674": "lxmert/modeling_lxmert.py:LxmertPredictionHeadTransform", "675": "lxmert/modeling_lxmert.py:LxmertLMPredictionHead", "676": "lxmert/modeling_lxmert.py:LxmertVisualAnswerHead", "677": "lxmert/modeling_lxmert.py:LxmertVisualObjHead", "678": "lxmert/modeling_lxmert.py:LxmertPreTrainingHeads", "679": "lxmert/modeling_lxmert.py:LxmertPreTrainedModel", "680": "lxmert/modeling_lxmert.py:LxmertModel", "681": "lxmert/modeling_lxmert.py:LxmertForPreTraining", "682": "lxmert/modeling_lxmert.py:LxmertForQuestionAnswering", "683": "mistral/modeling_mistral.py:MistralMLP", "684": "mistral/modeling_mistral.py:rotate_half", "685": "mistral/modeling_mistral.py:apply_rotary_pos_emb", "686": "mistral/modeling_mistral.py:repeat_kv", "687": "mistral/modeling_mistral.py:eager_attention_forward", "688": "mistral/modeling_mistral.py:MistralAttention", "689": "mistral/modeling_mistral.py:MistralRMSNorm", "690": "mistral/modeling_mistral.py:MistralDecoderLayer", "691": "mistral/modeling_mistral.py:MistralPreTrainedModel", "692": "mistral/modeling_mistral.py:MistralRotaryEmbedding", "693": "mistral/modeling_mistral.py:MistralModel", "694": "mistral/modeling_mistral.py:MistralForCausalLM", "695": "mistral/modeling_mistral.py:MistralForTokenClassification", "696": "mistral/modeling_mistral.py:MistralForSequenceClassification", "697": "mistral/modeling_mistral.py:MistralForQuestionAnswering", "698": "t5/modeling_t5.py:T5LayerNorm", "699": "t5/modeling_t5.py:T5DenseActDense", "700": "t5/modeling_t5.py:T5DenseGatedActDense", "701": "t5/modeling_t5.py:T5LayerFF", "702": "t5/modeling_t5.py:T5Attention", "703": "t5/modeling_t5.py:T5LayerSelfAttention", "704": "t5/modeling_t5.py:T5LayerCrossAttention", "705": "t5/modeling_t5.py:T5Block", "706": "t5/modeling_t5.py:T5ClassificationHead", "707": "t5/modeling_t5.py:T5PreTrainedModel", "708": "t5/modeling_t5.py:T5Stack", "709": "t5/modeling_t5.py:T5Model", "710": "t5/modeling_t5.py:T5ForConditionalGeneration", "711": "t5/modeling_t5.py:T5EncoderModel", "712": "t5/modeling_t5.py:T5ForSequenceClassification", "713": "t5/modeling_t5.py:T5ForTokenClassification", "714": "t5/modeling_t5.py:T5ForQuestionAnswering", "715": "rag/modeling_rag.py:RetrievAugLMMarginOutput", "716": "rag/modeling_rag.py:RetrievAugLMOutput", "717": "rag/modeling_rag.py:RagPreTrainedModel", "718": "rag/modeling_rag.py:RagModel", "719": "rag/modeling_rag.py:RagSequenceForGeneration", "720": "rag/modeling_rag.py:RagTokenForGeneration", "721": "gpt_neox/modeling_gpt_neox.py:GPTNeoXMLP", "722": "gpt_neox/modeling_gpt_neox.py:rotate_half", "723": "gpt_neox/modeling_gpt_neox.py:apply_rotary_pos_emb", "724": "gpt_neox/modeling_gpt_neox.py:eager_attention_forward", "725": "gpt_neox/modeling_gpt_neox.py:GPTNeoXAttention", "726": "gpt_neox/modeling_gpt_neox.py:GPTNeoXLayer", "727": "gpt_neox/modeling_gpt_neox.py:GPTNeoXRotaryEmbedding", "728": "gpt_neox/modeling_gpt_neox.py:GPTNeoXRMSNorm", "729": "gpt_neox/modeling_gpt_neox.py:GPTNeoXDecoderLayer", "730": "gpt_neox/modeling_gpt_neox.py:GPTNeoXPreTrainedModel", "731": "gpt_neox/modeling_gpt_neox.py:GPTNeoXModel", "732": "gpt_neox/modeling_gpt_neox.py:GPTNeoXForCausalLM", "733": "gpt_neox/modeling_gpt_neox.py:GPTNeoXForSequenceClassification", "734": "gpt_neox/modeling_gpt_neox.py:GPTNeoXForTokenClassification", "735": "gpt_neox/modeling_gpt_neox.py:GPTNeoXForQuestionAnswering", "736": "bigbird_pegasus/modeling_bigbird_pegasus.py:shift_tokens_right", "737": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusLearnedPositionalEmbedding", "738": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusScaledWordEmbedding", "739": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusSelfAttention", "740": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusBlockSparseAttention", "741": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusEncoderAttention", "742": "bigbird_pegasus/modeling_bigbird_pegasus.py:eager_attention_forward", "743": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusDecoderAttention", "744": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusEncoderLayer", "745": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusDecoderLayer", "746": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusClassificationHead", "747": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusPreTrainedModel", "748": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusEncoder", "749": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusDecoder", "750": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusModel", "751": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusForConditionalGeneration", "752": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusForSequenceClassification", "753": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusForQuestionAnswering", "754": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusDecoderWrapper", "755": "bigbird_pegasus/modeling_bigbird_pegasus.py:BigBirdPegasusForCausalLM", "756": "phi3/modeling_phi3.py:Phi3MLP", "757": "phi3/modeling_phi3.py:rotate_half", "758": "phi3/modeling_phi3.py:repeat_kv", "759": "phi3/modeling_phi3.py:eager_attention_forward", "760": "phi3/modeling_phi3.py:apply_rotary_pos_emb", "761": "phi3/modeling_phi3.py:Phi3Attention", "762": "phi3/modeling_phi3.py:Phi3RMSNorm", "763": "phi3/modeling_phi3.py:Phi3DecoderLayer", "764": "phi3/modeling_phi3.py:Phi3PreTrainedModel", "765": "phi3/modeling_phi3.py:Phi3RotaryEmbedding", "766": "phi3/modeling_phi3.py:Phi3Model", "767": "phi3/modeling_phi3.py:Phi3ForCausalLM", "768": "phi3/modeling_phi3.py:Phi3ForSequenceClassification", "769": "phi3/modeling_phi3.py:Phi3ForTokenClassification", "770": "unispeech/modeling_unispeech.py:UniSpeechForPreTrainingOutput", "771": "unispeech/modeling_unispeech.py:UniSpeechSamePadLayer", "772": "unispeech/modeling_unispeech.py:UniSpeechPositionalConvEmbedding", "773": "unispeech/modeling_unispeech.py:UniSpeechNoLayerNormConvLayer", "774": "unispeech/modeling_unispeech.py:UniSpeechLayerNormConvLayer", "775": "unispeech/modeling_unispeech.py:UniSpeechGroupNormConvLayer", "776": "unispeech/modeling_unispeech.py:UniSpeechFeatureEncoder", "777": "unispeech/modeling_unispeech.py:UniSpeechFeatureProjection", "778": "unispeech/modeling_unispeech.py:eager_attention_forward", "779": "unispeech/modeling_unispeech.py:UniSpeechAttention", "780": "unispeech/modeling_unispeech.py:UniSpeechFeedForward", "781": "unispeech/modeling_unispeech.py:UniSpeechEncoderLayer", "782": "unispeech/modeling_unispeech.py:UniSpeechEncoder", "783": "unispeech/modeling_unispeech.py:UniSpeechAttnAdapterLayer", "784": "unispeech/modeling_unispeech.py:UniSpeechEncoderLayerStableLayerNorm", "785": "unispeech/modeling_unispeech.py:UniSpeechEncoderStableLayerNorm", "786": "unispeech/modeling_unispeech.py:UniSpeechGumbelVectorQuantizer", "787": "unispeech/modeling_unispeech.py:UniSpeechPreTrainedModel", "788": "unispeech/modeling_unispeech.py:_compute_mask_indices", "789": "unispeech/modeling_unispeech.py:UniSpeechModel", "790": "unispeech/modeling_unispeech.py:UniSpeechForPreTraining", "791": "unispeech/modeling_unispeech.py:UniSpeechForCTC", "792": "unispeech/modeling_unispeech.py:UniSpeechForSequenceClassification", "793": "olmo/modeling_olmo.py:OlmoLayerNorm", "794": "olmo/modeling_olmo.py:OlmoMLP", "795": "olmo/modeling_olmo.py:rotate_half", "796": "olmo/modeling_olmo.py:repeat_kv", "797": "olmo/modeling_olmo.py:eager_attention_forward", "798": "olmo/modeling_olmo.py:apply_rotary_pos_emb", "799": "olmo/modeling_olmo.py:OlmoAttention", "800": "olmo/modeling_olmo.py:OlmoDecoderLayer", "801": "olmo/modeling_olmo.py:OlmoRotaryEmbedding", "802": "olmo/modeling_olmo.py:OlmoPreTrainedModel", "803": "olmo/modeling_olmo.py:OlmoModel", "804": "olmo/modeling_olmo.py:OlmoForCausalLM", "805": "led/modeling_led.py:shift_tokens_right", "806": "led/modeling_led.py:_prepare_4d_attention_mask_inverted", "807": "led/modeling_led.py:LEDLearnedPositionalEmbedding", "808": "led/modeling_led.py:LEDEncoderSelfAttention", "809": "led/modeling_led.py:LEDEncoderAttention", "810": "led/modeling_led.py:LEDDecoderAttention", "811": "led/modeling_led.py:LEDEncoderLayer", "812": "led/modeling_led.py:LEDDecoderLayer", "813": "led/modeling_led.py:LEDClassificationHead", "814": "led/modeling_led.py:LEDPreTrainedModel", "815": "led/modeling_led.py:LEDEncoderBaseModelOutput", "816": "led/modeling_led.py:LEDSeq2SeqModelOutput", "817": "led/modeling_led.py:LEDSeq2SeqLMOutput", "818": "led/modeling_led.py:LEDSeq2SeqSequenceClassifierOutput", "819": "led/modeling_led.py:LEDSeq2SeqQuestionAnsweringModelOutput", "820": "led/modeling_led.py:LEDEncoder", "821": "led/modeling_led.py:LEDDecoder", "822": "led/modeling_led.py:LEDModel", "823": "led/modeling_led.py:LEDForConditionalGeneration", "824": "led/modeling_led.py:LEDForSequenceClassification", "825": "led/modeling_led.py:LEDForQuestionAnswering", "826": "bloom/modeling_bloom.py:build_alibi_tensor", "827": "bloom/modeling_bloom.py:dropout_add", "828": "bloom/modeling_bloom.py:bloom_gelu_forward", "829": "bloom/modeling_bloom.py:bloom_gelu_back", "830": "bloom/modeling_bloom.py:GeLUFunction", "831": "bloom/modeling_bloom.py:BloomGelu", "832": "bloom/modeling_bloom.py:BloomAttention", "833": "bloom/modeling_bloom.py:BloomMLP", "834": "bloom/modeling_bloom.py:BloomBlock", "835": "bloom/modeling_bloom.py:BloomPreTrainedModel", "836": "bloom/modeling_bloom.py:BloomModel", "837": "bloom/modeling_bloom.py:BloomForCausalLM", "838": "bloom/modeling_bloom.py:BloomForSequenceClassification", "839": "bloom/modeling_bloom.py:BloomForTokenClassification", "840": "bloom/modeling_bloom.py:BloomForQuestionAnswering", "841": "helium/modeling_helium.py:HeliumRMSNorm", "842": "helium/modeling_helium.py:HeliumRotaryEmbedding", "843": "helium/modeling_helium.py:HeliumMLP", "844": "helium/modeling_helium.py:repeat_kv", "845": "helium/modeling_helium.py:eager_attention_forward", "846": "helium/modeling_helium.py:rotate_half", "847": "helium/modeling_helium.py:apply_rotary_pos_emb", "848": "helium/modeling_helium.py:HeliumAttention", "849": "helium/modeling_helium.py:HeliumDecoderLayer", "850": "helium/modeling_helium.py:HeliumPreTrainedModel", "851": "helium/modeling_helium.py:HeliumModel", "852": "helium/modeling_helium.py:HeliumForCausalLM", "853": "helium/modeling_helium.py:HeliumForSequenceClassification", "854": "helium/modeling_helium.py:HeliumForTokenClassification", "855": "musicgen/modeling_musicgen.py:MusicgenUnconditionalInput", "856": "musicgen/modeling_musicgen.py:shift_tokens_right", "857": "musicgen/modeling_musicgen.py:MusicgenSinusoidalPositionalEmbedding", "858": "musicgen/modeling_musicgen.py:eager_attention_forward", "859": "musicgen/modeling_musicgen.py:MusicgenAttention", "860": "musicgen/modeling_musicgen.py:MusicgenDecoderLayer", "861": "musicgen/modeling_musicgen.py:MusicgenPreTrainedModel", "862": "musicgen/modeling_musicgen.py:MusicgenDecoder", "863": "musicgen/modeling_musicgen.py:MusicgenModel", "864": "musicgen/modeling_musicgen.py:MusicgenForCausalLM", "865": "musicgen/modeling_musicgen.py:MusicgenForConditionalGeneration", "866": "roc_bert/modeling_roc_bert.py:RoCBertEmbeddings", "867": "roc_bert/modeling_roc_bert.py:eager_attention_forward", "868": "roc_bert/modeling_roc_bert.py:RoCBertSelfAttention", "869": "roc_bert/modeling_roc_bert.py:RoCBertCrossAttention", "870": "roc_bert/modeling_roc_bert.py:RoCBertSelfOutput", "871": "roc_bert/modeling_roc_bert.py:RoCBertAttention", "872": "roc_bert/modeling_roc_bert.py:RoCBertIntermediate", "873": "roc_bert/modeling_roc_bert.py:RoCBertOutput", "874": "roc_bert/modeling_roc_bert.py:RoCBertLayer", "875": "roc_bert/modeling_roc_bert.py:RoCBertEncoder", "876": "roc_bert/modeling_roc_bert.py:RoCBertPooler", "877": "roc_bert/modeling_roc_bert.py:RoCBertPredictionHeadTransform", "878": "roc_bert/modeling_roc_bert.py:RoCBertLMPredictionHead", "879": "roc_bert/modeling_roc_bert.py:RoCBertOnlyMLMHead", "880": "roc_bert/modeling_roc_bert.py:RoCBertPreTrainedModel", "881": "roc_bert/modeling_roc_bert.py:RoCBertModel", "882": "roc_bert/modeling_roc_bert.py:RoCBertForPreTraining", "883": "roc_bert/modeling_roc_bert.py:RoCBertForMaskedLM", "884": "roc_bert/modeling_roc_bert.py:RoCBertForCausalLM", "885": "roc_bert/modeling_roc_bert.py:RoCBertForSequenceClassification", "886": "roc_bert/modeling_roc_bert.py:RoCBertForMultipleChoice", "887": "roc_bert/modeling_roc_bert.py:RoCBertForTokenClassification", "888": "roc_bert/modeling_roc_bert.py:RoCBertForQuestionAnswering", "889": "bitnet/modeling_bitnet.py:BitNetRMSNorm", "890": "bitnet/modeling_bitnet.py:BitNetMLP", "891": "bitnet/modeling_bitnet.py:rotate_half", "892": "bitnet/modeling_bitnet.py:apply_rotary_pos_emb", "893": "bitnet/modeling_bitnet.py:repeat_kv", "894": "bitnet/modeling_bitnet.py:eager_attention_forward", "895": "bitnet/modeling_bitnet.py:BitNetAttention", "896": "bitnet/modeling_bitnet.py:BitNetDecoderLayer", "897": "bitnet/modeling_bitnet.py:BitNetRotaryEmbedding", "898": "bitnet/modeling_bitnet.py:BitNetPreTrainedModel", "899": "bitnet/modeling_bitnet.py:BitNetModel", "900": "bitnet/modeling_bitnet.py:BitNetForCausalLM", "901": "mask2former/modeling_mask2former.py:Mask2FormerPixelDecoderOutput", "902": "mask2former/modeling_mask2former.py:Mask2FormerMaskedAttentionDecoderOutput", "903": "mask2former/modeling_mask2former.py:Mask2FormerPixelLevelModuleOutput", "904": "mask2former/modeling_mask2former.py:Mask2FormerModelOutput", "905": "mask2former/modeling_mask2former.py:Mask2FormerForUniversalSegmentationOutput", "906": "mask2former/modeling_mask2former.py:sample_point", "907": "mask2former/modeling_mask2former.py:dice_loss", "908": "mask2former/modeling_mask2former.py:sigmoid_cross_entropy_loss", "909": "mask2former/modeling_mask2former.py:pair_wise_dice_loss", "910": "mask2former/modeling_mask2former.py:pair_wise_sigmoid_cross_entropy_loss", "911": "mask2former/modeling_mask2former.py:Mask2FormerHungarianMatcher", "912": "mask2former/modeling_mask2former.py:Mask2FormerLoss", "913": "mask2former/modeling_mask2former.py:multi_scale_deformable_attention", "914": "mask2former/modeling_mask2former.py:Mask2FormerSinePositionEmbedding", "915": "mask2former/modeling_mask2former.py:Mask2FormerPixelDecoderEncoderMultiscaleDeformableAttention", "916": "mask2former/modeling_mask2former.py:Mask2FormerPixelDecoderEncoderLayer", "917": "mask2former/modeling_mask2former.py:Mask2FormerPixelDecoderEncoderOnly", "918": "mask2former/modeling_mask2former.py:Mask2FormerPixelDecoder", "919": "mask2former/modeling_mask2former.py:Mask2FormerPixelLevelModule", "920": "mask2former/modeling_mask2former.py:Mask2FormerAttention", "921": "mask2former/modeling_mask2former.py:Mask2FormerMaskedAttentionDecoderLayer", "922": "mask2former/modeling_mask2former.py:Mask2FormerMaskedAttentionDecoder", "923": "mask2former/modeling_mask2former.py:Mask2FormerPredictionBlock", "924": "mask2former/modeling_mask2former.py:Mask2FormerMLPPredictionHead", "925": "mask2former/modeling_mask2former.py:Mask2FormerMaskPredictor", "926": "mask2former/modeling_mask2former.py:Mask2FormerTransformerModule", "927": "mask2former/modeling_mask2former.py:Mask2FormerPreTrainedModel", "928": "mask2former/modeling_mask2former.py:Mask2FormerModel", "929": "mask2former/modeling_mask2former.py:Mask2FormerForUniversalSegmentation", "930": "granitemoe/modeling_granitemoe.py:load_balancing_loss_func", "931": "granitemoe/modeling_granitemoe.py:GraniteMoeRMSNorm", "932": "granitemoe/modeling_granitemoe.py:GraniteMoeRotaryEmbedding", "933": "granitemoe/modeling_granitemoe.py:rotate_half", "934": "granitemoe/modeling_granitemoe.py:apply_rotary_pos_emb", "935": "granitemoe/modeling_granitemoe.py:GraniteMoeParallelExperts", "936": "granitemoe/modeling_granitemoe.py:GraniteMoeTopKGating", "937": "granitemoe/modeling_granitemoe.py:GraniteMoeMoE", "938": "granitemoe/modeling_granitemoe.py:repeat_kv", "939": "granitemoe/modeling_granitemoe.py:GraniteMoeAttention", "940": "granitemoe/modeling_granitemoe.py:eager_attention_forward", "941": "granitemoe/modeling_granitemoe.py:GraniteMoeDecoderLayer", "942": "granitemoe/modeling_granitemoe.py:GraniteMoePreTrainedModel", "943": "granitemoe/modeling_granitemoe.py:GraniteMoeModel", "944": "granitemoe/modeling_granitemoe.py:GraniteMoeForCausalLM", "945": "falcon_h1/modeling_falcon_h1.py:FalconHybridMambaAttentionDynamicCache", "946": "falcon_h1/modeling_falcon_h1.py:FalconH1RotaryEmbedding", "947": "falcon_h1/modeling_falcon_h1.py:rotate_half", "948": "falcon_h1/modeling_falcon_h1.py:apply_rotary_pos_emb", "949": "falcon_h1/modeling_falcon_h1.py:repeat_kv", "950": "falcon_h1/modeling_falcon_h1.py:eager_attention_forward", "951": "falcon_h1/modeling_falcon_h1.py:FalconH1Attention", "952": "falcon_h1/modeling_falcon_h1.py:FalconH1RMSNormGated", "953": "falcon_h1/modeling_falcon_h1.py:pad_tensor_by_size", "954": "falcon_h1/modeling_falcon_h1.py:reshape_into_chunks", "955": "falcon_h1/modeling_falcon_h1.py:segment_sum", "956": "falcon_h1/modeling_falcon_h1.py:apply_mask_to_padding_states", "957": "falcon_h1/modeling_falcon_h1.py:FalconH1Mixer", "958": "falcon_h1/modeling_falcon_h1.py:FalconH1MLP", "959": "falcon_h1/modeling_falcon_h1.py:FalconH1RMSNorm", "960": "falcon_h1/modeling_falcon_h1.py:FalconH1DecoderLayer", "961": "falcon_h1/modeling_falcon_h1.py:FalconH1PreTrainedModel", "962": "falcon_h1/modeling_falcon_h1.py:compute_mup_vector", "963": "falcon_h1/modeling_falcon_h1.py:FalconH1Model", "964": "falcon_h1/modeling_falcon_h1.py:FalconH1ForCausalLM", "965": "table_transformer/modeling_table_transformer.py:TableTransformerDecoderOutput", "966": "table_transformer/modeling_table_transformer.py:TableTransformerModelOutput", "967": "table_transformer/modeling_table_transformer.py:TableTransformerObjectDetectionOutput", "968": "table_transformer/modeling_table_transformer.py:TableTransformerFrozenBatchNorm2d", "969": "table_transformer/modeling_table_transformer.py:replace_batch_norm", "970": "table_transformer/modeling_table_transformer.py:TableTransformerConvEncoder", "971": "table_transformer/modeling_table_transformer.py:TableTransformerConvModel", "972": "table_transformer/modeling_table_transformer.py:TableTransformerSinePositionEmbedding", "973": "table_transformer/modeling_table_transformer.py:TableTransformerLearnedPositionEmbedding", "974": "table_transformer/modeling_table_transformer.py:build_position_encoding", "975": "table_transformer/modeling_table_transformer.py:TableTransformerAttention", "976": "table_transformer/modeling_table_transformer.py:TableTransformerEncoderLayer", "977": "table_transformer/modeling_table_transformer.py:TableTransformerDecoderLayer", "978": "table_transformer/modeling_table_transformer.py:TableTransformerPreTrainedModel", "979": "table_transformer/modeling_table_transformer.py:TableTransformerEncoder", "980": "table_transformer/modeling_table_transformer.py:TableTransformerDecoder", "981": "table_transformer/modeling_table_transformer.py:TableTransformerModel", "982": "table_transformer/modeling_table_transformer.py:TableTransformerForObjectDetection", "983": "table_transformer/modeling_table_transformer.py:TableTransformerMLPPredictionHead", "984": "speecht5/modeling_speecht5.py:shift_tokens_right", "985": "speecht5/modeling_speecht5.py:shift_spectrograms_right", "986": "speecht5/modeling_speecht5.py:_compute_mask_indices", "987": "speecht5/modeling_speecht5.py:SpeechT5NoLayerNormConvLayer", "988": "speecht5/modeling_speecht5.py:SpeechT5LayerNormConvLayer", "989": "speecht5/modeling_speecht5.py:SpeechT5GroupNormConvLayer", "990": "speecht5/modeling_speecht5.py:SpeechT5SinusoidalPositionalEmbedding", "991": "speecht5/modeling_speecht5.py:SpeechT5PositionalConvEmbedding", "992": "speecht5/modeling_speecht5.py:SpeechT5ScaledPositionalEncoding", "993": "speecht5/modeling_speecht5.py:SpeechT5RelativePositionalEncoding", "994": "speecht5/modeling_speecht5.py:SpeechT5SamePadLayer", "995": "speecht5/modeling_speecht5.py:SpeechT5FeatureEncoder", "996": "speecht5/modeling_speecht5.py:SpeechT5FeatureProjection", "997": "speecht5/modeling_speecht5.py:SpeechT5SpeechEncoderPrenet", "998": "speecht5/modeling_speecht5.py:SpeechT5SpeechDecoderPrenet", "999": "speecht5/modeling_speecht5.py:SpeechT5BatchNormConvLayer", "1000": "speecht5/modeling_speecht5.py:SpeechT5SpeechDecoderPostnet", "1001": "speecht5/modeling_speecht5.py:SpeechT5TextEncoderPrenet", "1002": "speecht5/modeling_speecht5.py:SpeechT5TextDecoderPrenet", "1003": "speecht5/modeling_speecht5.py:SpeechT5TextDecoderPostnet", "1004": "speecht5/modeling_speecht5.py:SpeechT5Attention", "1005": "speecht5/modeling_speecht5.py:SpeechT5FeedForward", "1006": "speecht5/modeling_speecht5.py:SpeechT5EncoderLayer", "1007": "speecht5/modeling_speecht5.py:SpeechT5DecoderLayer", "1008": "speecht5/modeling_speecht5.py:SpeechT5PreTrainedModel", "1009": "speecht5/modeling_speecht5.py:SpeechT5Encoder", "1010": "speecht5/modeling_speecht5.py:SpeechT5EncoderWithSpeechPrenet", "1011": "speecht5/modeling_speecht5.py:SpeechT5EncoderWithTextPrenet", "1012": "speecht5/modeling_speecht5.py:SpeechT5EncoderWithoutPrenet", "1013": "speecht5/modeling_speecht5.py:SpeechT5Decoder", "1014": "speecht5/modeling_speecht5.py:SpeechT5DecoderWithSpeechPrenet", "1015": "speecht5/modeling_speecht5.py:SpeechT5DecoderWithTextPrenet", "1016": "speecht5/modeling_speecht5.py:SpeechT5DecoderWithoutPrenet", "1017": "speecht5/modeling_speecht5.py:SpeechT5GuidedMultiheadAttentionLoss", "1018": "speecht5/modeling_speecht5.py:SpeechT5SpectrogramLoss", "1019": "speecht5/modeling_speecht5.py:SpeechT5Model", "1020": "speecht5/modeling_speecht5.py:SpeechT5ForSpeechToText", "1021": "speecht5/modeling_speecht5.py:_generate_speech", "1022": "speecht5/modeling_speecht5.py:SpeechT5ForTextToSpeech", "1023": "speecht5/modeling_speecht5.py:SpeechT5ForSpeechToSpeech", "1024": "speecht5/modeling_speecht5.py:HifiGanResidualBlock", "1025": "speecht5/modeling_speecht5.py:SpeechT5HifiGan", "1026": "hiera/modeling_hiera.py:HieraEncoderOutput", "1027": "hiera/modeling_hiera.py:HieraModelOutput", "1028": "hiera/modeling_hiera.py:HieraForImageClassificationOutput", "1029": "hiera/modeling_hiera.py:HieraForPreTrainingOutput", "1030": "hiera/modeling_hiera.py:HieraPatchEmbeddings", "1031": "hiera/modeling_hiera.py:HieraEmbeddings", "1032": "hiera/modeling_hiera.py:HieraMaskUnitAttention", "1033": "hiera/modeling_hiera.py:drop_path", "1034": "hiera/modeling_hiera.py:HieraDropPath", "1035": "hiera/modeling_hiera.py:HieraMlp", "1036": "hiera/modeling_hiera.py:HieraLayer", "1037": "hiera/modeling_hiera.py:HieraStage", "1038": "hiera/modeling_hiera.py:undo_windowing", "1039": "hiera/modeling_hiera.py:HieraEncoder", "1040": "hiera/modeling_hiera.py:unroll", "1041": "hiera/modeling_hiera.py:HieraPreTrainedModel", "1042": "hiera/modeling_hiera.py:HieraPooler", "1043": "hiera/modeling_hiera.py:HieraModel", "1044": "hiera/modeling_hiera.py:HieraDecoder", "1045": "hiera/modeling_hiera.py:HieraMultiScaleHead", "1046": "hiera/modeling_hiera.py:HieraForPreTraining", "1047": "hiera/modeling_hiera.py:HieraForImageClassification", "1048": "hiera/modeling_hiera.py:HieraBackbone", "1049": "canine/modeling_canine.py:CanineModelOutputWithPooling", "1050": "canine/modeling_canine.py:CanineEmbeddings", "1051": "canine/modeling_canine.py:CharactersToMolecules", "1052": "canine/modeling_canine.py:ConvProjection", "1053": "canine/modeling_canine.py:CanineSelfAttention", "1054": "canine/modeling_canine.py:CanineSelfOutput", "1055": "canine/modeling_canine.py:CanineAttention", "1056": "canine/modeling_canine.py:CanineIntermediate", "1057": "canine/modeling_canine.py:CanineOutput", "1058": "canine/modeling_canine.py:CanineLayer", "1059": "canine/modeling_canine.py:CanineEncoder", "1060": "canine/modeling_canine.py:CaninePooler", "1061": "canine/modeling_canine.py:CaninePredictionHeadTransform", "1062": "canine/modeling_canine.py:CanineLMPredictionHead", "1063": "canine/modeling_canine.py:CanineOnlyMLMHead", "1064": "canine/modeling_canine.py:CaninePreTrainedModel", "1065": "canine/modeling_canine.py:CanineModel", "1066": "canine/modeling_canine.py:CanineForSequenceClassification", "1067": "canine/modeling_canine.py:CanineForMultipleChoice", "1068": "canine/modeling_canine.py:CanineForTokenClassification", "1069": "canine/modeling_canine.py:CanineForQuestionAnswering", "1070": "xlm_roberta/modeling_xlm_roberta.py:eager_attention_forward", "1071": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaSelfAttention", "1072": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaCrossAttention", "1073": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaSelfOutput", "1074": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaAttention", "1075": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaIntermediate", "1076": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaOutput", "1077": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaLayer", "1078": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaLMHead", "1079": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaPreTrainedModel", "1080": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaEmbeddings", "1081": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaEncoder", "1082": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaPooler", "1083": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaModel", "1084": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaForCausalLM", "1085": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaForMaskedLM", "1086": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaClassificationHead", "1087": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaForSequenceClassification", "1088": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaForMultipleChoice", "1089": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaForTokenClassification", "1090": "xlm_roberta/modeling_xlm_roberta.py:XLMRobertaForQuestionAnswering", "1091": "zoedepth/modeling_zoedepth.py:ZoeDepthDepthEstimatorOutput", "1092": "zoedepth/modeling_zoedepth.py:ZoeDepthReassembleStage", "1093": "zoedepth/modeling_zoedepth.py:ZoeDepthReassembleLayer", "1094": "zoedepth/modeling_zoedepth.py:ZoeDepthFeatureFusionStage", "1095": "zoedepth/modeling_zoedepth.py:ZoeDepthPreActResidualLayer", "1096": "zoedepth/modeling_zoedepth.py:ZoeDepthFeatureFusionLayer", "1097": "zoedepth/modeling_zoedepth.py:ZoeDepthNeck", "1098": "zoedepth/modeling_zoedepth.py:ZoeDepthRelativeDepthEstimationHead", "1099": "zoedepth/modeling_zoedepth.py:log_binom", "1100": "zoedepth/modeling_zoedepth.py:LogBinomialSoftmax", "1101": "zoedepth/modeling_zoedepth.py:ZoeDepthConditionalLogBinomialSoftmax", "1102": "zoedepth/modeling_zoedepth.py:ZoeDepthSeedBinRegressor", "1103": "zoedepth/modeling_zoedepth.py:inv_attractor", "1104": "zoedepth/modeling_zoedepth.py:ZoeDepthAttractorLayer", "1105": "zoedepth/modeling_zoedepth.py:ZoeDepthAttractorLayerUnnormed", "1106": "zoedepth/modeling_zoedepth.py:ZoeDepthProjector", "1107": "zoedepth/modeling_zoedepth.py:ZoeDepthMultiheadAttention", "1108": "zoedepth/modeling_zoedepth.py:ZoeDepthTransformerEncoderLayer", "1109": "zoedepth/modeling_zoedepth.py:ZoeDepthPatchTransformerEncoder", "1110": "zoedepth/modeling_zoedepth.py:ZoeDepthMLPClassifier", "1111": "zoedepth/modeling_zoedepth.py:ZoeDepthMultipleMetricDepthEstimationHeads", "1112": "zoedepth/modeling_zoedepth.py:ZoeDepthMetricDepthEstimationHead", "1113": "zoedepth/modeling_zoedepth.py:ZoeDepthPreTrainedModel", "1114": "zoedepth/modeling_zoedepth.py:ZoeDepthForDepthEstimation", "1115": "groupvit/modeling_groupvit.py:contrastive_loss", "1116": "groupvit/modeling_groupvit.py:groupvit_loss", "1117": "groupvit/modeling_groupvit.py:hard_softmax", "1118": "groupvit/modeling_groupvit.py:gumbel_softmax", "1119": "groupvit/modeling_groupvit.py:resize_attention_map", "1120": "groupvit/modeling_groupvit.py:get_grouping_from_attentions", "1121": "groupvit/modeling_groupvit.py:GroupViTCrossAttentionLayer", "1122": "groupvit/modeling_groupvit.py:GroupViTAssignAttention", "1123": "groupvit/modeling_groupvit.py:GroupViTTokenAssign", "1124": "groupvit/modeling_groupvit.py:GroupViTModelOutput", "1125": "groupvit/modeling_groupvit.py:GroupViTPatchEmbeddings", "1126": "groupvit/modeling_groupvit.py:GroupViTVisionEmbeddings", "1127": "groupvit/modeling_groupvit.py:GroupViTTextEmbeddings", "1128": "groupvit/modeling_groupvit.py:GroupViTStage", "1129": "groupvit/modeling_groupvit.py:GroupViTMLP", "1130": "groupvit/modeling_groupvit.py:GroupViTMixerMLP", "1131": "groupvit/modeling_groupvit.py:GroupViTAttention", "1132": "groupvit/modeling_groupvit.py:GroupViTEncoderLayer", "1133": "groupvit/modeling_groupvit.py:GroupViTPreTrainedModel", "1134": "groupvit/modeling_groupvit.py:GroupViTVisionEncoder", "1135": "groupvit/modeling_groupvit.py:GroupViTTextEncoder", "1136": "groupvit/modeling_groupvit.py:GroupViTTextTransformer", "1137": "groupvit/modeling_groupvit.py:GroupViTTextModel", "1138": "groupvit/modeling_groupvit.py:GroupViTVisionTransformer", "1139": "groupvit/modeling_groupvit.py:GroupViTVisionModel", "1140": "groupvit/modeling_groupvit.py:GroupViTModel", "1141": "mt5/modeling_mt5.py:MT5LayerNorm", "1142": "mt5/modeling_mt5.py:MT5DenseActDense", "1143": "mt5/modeling_mt5.py:MT5DenseGatedActDense", "1144": "mt5/modeling_mt5.py:MT5LayerFF", "1145": "mt5/modeling_mt5.py:MT5Attention", "1146": "mt5/modeling_mt5.py:MT5LayerSelfAttention", "1147": "mt5/modeling_mt5.py:MT5LayerCrossAttention", "1148": "mt5/modeling_mt5.py:MT5Block", "1149": "mt5/modeling_mt5.py:MT5ClassificationHead", "1150": "mt5/modeling_mt5.py:MT5PreTrainedModel", "1151": "mt5/modeling_mt5.py:MT5Stack", "1152": "mt5/modeling_mt5.py:MT5Model", "1153": "mt5/modeling_mt5.py:MT5ForConditionalGeneration", "1154": "mt5/modeling_mt5.py:MT5EncoderModel", "1155": "mt5/modeling_mt5.py:MT5ForSequenceClassification", "1156": "mt5/modeling_mt5.py:MT5ForTokenClassification", "1157": "mt5/modeling_mt5.py:MT5ForQuestionAnswering", "1158": "mgp_str/modeling_mgp_str.py:drop_path", "1159": "mgp_str/modeling_mgp_str.py:MgpstrDropPath", "1160": "mgp_str/modeling_mgp_str.py:MgpstrModelOutput", "1161": "mgp_str/modeling_mgp_str.py:MgpstrEmbeddings", "1162": "mgp_str/modeling_mgp_str.py:MgpstrMlp", "1163": "mgp_str/modeling_mgp_str.py:MgpstrAttention", "1164": "mgp_str/modeling_mgp_str.py:MgpstrLayer", "1165": "mgp_str/modeling_mgp_str.py:MgpstrEncoder", "1166": "mgp_str/modeling_mgp_str.py:MgpstrA3Module", "1167": "mgp_str/modeling_mgp_str.py:MgpstrPreTrainedModel", "1168": "mgp_str/modeling_mgp_str.py:MgpstrModel", "1169": "mgp_str/modeling_mgp_str.py:MgpstrForSceneTextRecognition", "1170": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Embeddings", "1171": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2SelfAttention", "1172": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Attention", "1173": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2SelfOutput", "1174": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Intermediate", "1175": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Output", "1176": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Layer", "1177": "layoutlmv2/modeling_layoutlmv2.py:relative_position_bucket", "1178": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Encoder", "1179": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2PreTrainedModel", "1180": "layoutlmv2/modeling_layoutlmv2.py:my_convert_sync_batchnorm", "1181": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2VisualBackbone", "1182": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Pooler", "1183": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2Model", "1184": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2ForSequenceClassification", "1185": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2ForTokenClassification", "1186": "layoutlmv2/modeling_layoutlmv2.py:LayoutLMv2ForQuestionAnswering", "1187": "mllama/modeling_mllama.py:_prepare_cross_attention_mask", "1188": "mllama/modeling_mllama.py:_prepare_aspect_ratio_attention_mask", "1189": "mllama/modeling_mllama.py:MllamaPrecomputedAspectRatioEmbedding", "1190": "mllama/modeling_mllama.py:MllamaPrecomputedPositionEmbedding", "1191": "mllama/modeling_mllama.py:MllamaVisionMLP", "1192": "mllama/modeling_mllama.py:repeat_kv", "1193": "mllama/modeling_mllama.py:eager_attention_forward", "1194": "mllama/modeling_mllama.py:MllamaVisionAttention", "1195": "mllama/modeling_mllama.py:MllamaVisionEncoderLayer", "1196": "mllama/modeling_mllama.py:MllamaVisionEncoder", "1197": "mllama/modeling_mllama.py:MllamaTextRMSNorm", "1198": "mllama/modeling_mllama.py:MllamaTextCrossAttention", "1199": "mllama/modeling_mllama.py:rotate_half", "1200": "mllama/modeling_mllama.py:apply_rotary_pos_emb", "1201": "mllama/modeling_mllama.py:MllamaTextSelfAttention", "1202": "mllama/modeling_mllama.py:MllamaTextMLP", "1203": "mllama/modeling_mllama.py:MllamaSelfAttentionDecoderLayer", "1204": "mllama/modeling_mllama.py:MllamaCrossAttentionDecoderLayer", "1205": "mllama/modeling_mllama.py:MllamaRotaryEmbedding", "1206": "mllama/modeling_mllama.py:MllamaPreTrainedModel", "1207": "mllama/modeling_mllama.py:MllamaVisionModel", "1208": "mllama/modeling_mllama.py:MllamaTextModel", "1209": "mllama/modeling_mllama.py:MllamaForCausalLM", "1210": "mllama/modeling_mllama.py:MllamaModel", "1211": "mllama/modeling_mllama.py:MllamaForConditionalGeneration", "1212": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinModelOutputWithPooling", "1213": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinBaseModelOutput", "1214": "maskformer/modeling_maskformer_swin.py:window_partition", "1215": "maskformer/modeling_maskformer_swin.py:window_reverse", "1216": "maskformer/modeling_maskformer_swin.py:drop_path", "1217": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinEmbeddings", "1218": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinPatchEmbeddings", "1219": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinPatchMerging", "1220": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinDropPath", "1221": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinSelfAttention", "1222": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinSelfOutput", "1223": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinAttention", "1224": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinIntermediate", "1225": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinOutput", "1226": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinLayer", "1227": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinStage", "1228": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinEncoder", "1229": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinPreTrainedModel", "1230": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinModel", "1231": "maskformer/modeling_maskformer_swin.py:MaskFormerSwinBackbone", "1232": "maskformer/modeling_maskformer.py:DetrDecoderOutput", "1233": "maskformer/modeling_maskformer.py:MaskFormerPixelLevelModuleOutput", "1234": "maskformer/modeling_maskformer.py:MaskFormerPixelDecoderOutput", "1235": "maskformer/modeling_maskformer.py:MaskFormerModelOutput", "1236": "maskformer/modeling_maskformer.py:MaskFormerForInstanceSegmentationOutput", "1237": "maskformer/modeling_maskformer.py:upsample_like", "1238": "maskformer/modeling_maskformer.py:dice_loss", "1239": "maskformer/modeling_maskformer.py:sigmoid_focal_loss", "1240": "maskformer/modeling_maskformer.py:pair_wise_dice_loss", "1241": "maskformer/modeling_maskformer.py:pair_wise_sigmoid_focal_loss", "1242": "maskformer/modeling_maskformer.py:DetrAttention", "1243": "maskformer/modeling_maskformer.py:DetrDecoderLayer", "1244": "maskformer/modeling_maskformer.py:DetrDecoder", "1245": "maskformer/modeling_maskformer.py:MaskFormerHungarianMatcher", "1246": "maskformer/modeling_maskformer.py:MaskFormerLoss", "1247": "maskformer/modeling_maskformer.py:MaskFormerFPNConvLayer", "1248": "maskformer/modeling_maskformer.py:MaskFormerFPNLayer", "1249": "maskformer/modeling_maskformer.py:MaskFormerFPNModel", "1250": "maskformer/modeling_maskformer.py:MaskFormerPixelDecoder", "1251": "maskformer/modeling_maskformer.py:MaskFormerSinePositionEmbedding", "1252": "maskformer/modeling_maskformer.py:PredictionBlock", "1253": "maskformer/modeling_maskformer.py:MaskformerMLPPredictionHead", "1254": "maskformer/modeling_maskformer.py:MaskFormerPixelLevelModule", "1255": "maskformer/modeling_maskformer.py:MaskFormerTransformerModule", "1256": "maskformer/modeling_maskformer.py:MaskFormerPreTrainedModel", "1257": "maskformer/modeling_maskformer.py:MaskFormerModel", "1258": "maskformer/modeling_maskformer.py:MaskFormerForInstanceSegmentation", "1259": "blenderbot_small/modeling_blenderbot_small.py:shift_tokens_right", "1260": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallLearnedPositionalEmbedding", "1261": "blenderbot_small/modeling_blenderbot_small.py:eager_attention_forward", "1262": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallAttention", "1263": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallEncoderLayer", "1264": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallDecoderLayer", "1265": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallPreTrainedModel", "1266": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallEncoder", "1267": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallDecoder", "1268": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallModel", "1269": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallForConditionalGeneration", "1270": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallDecoderWrapper", "1271": "blenderbot_small/modeling_blenderbot_small.py:BlenderbotSmallForCausalLM", "1272": "got_ocr2/modeling_got_ocr2.py:GotOcr2MLPBlock", "1273": "got_ocr2/modeling_got_ocr2.py:GotOcr2VisionAttention", "1274": "got_ocr2/modeling_got_ocr2.py:GotOcr2VisionLayer", "1275": "got_ocr2/modeling_got_ocr2.py:GotOcr2PreTrainedModel", "1276": "got_ocr2/modeling_got_ocr2.py:GotOcr2VisionEncoderOutput", "1277": "got_ocr2/modeling_got_ocr2.py:GotOcr2PatchEmbeddings", "1278": "got_ocr2/modeling_got_ocr2.py:GotOcr2LayerNorm", "1279": "got_ocr2/modeling_got_ocr2.py:GotOcr2VisionNeck", "1280": "got_ocr2/modeling_got_ocr2.py:GotOcr2VisionEncoder", "1281": "got_ocr2/modeling_got_ocr2.py:GotOcr2MultiModalProjector", "1282": "got_ocr2/modeling_got_ocr2.py:GotOcr2CausalLMOutputWithPast", "1283": "got_ocr2/modeling_got_ocr2.py:GotOcr2ModelOutputWithPast", "1284": "got_ocr2/modeling_got_ocr2.py:GotOcr2Model", "1285": "got_ocr2/modeling_got_ocr2.py:GotOcr2ForConditionalGeneration", "1286": "vjepa2/modeling_vjepa2.py:VJEPA2WithMaskedInputPredictorOutput", "1287": "vjepa2/modeling_vjepa2.py:VJEPA2WithMaskedInputModelOutput", "1288": "vjepa2/modeling_vjepa2.py:VJEPA2PatchEmbeddings3D", "1289": "vjepa2/modeling_vjepa2.py:VJEPA2Embeddings", "1290": "vjepa2/modeling_vjepa2.py:eager_attention_forward", "1291": "vjepa2/modeling_vjepa2.py:rotate_queries_or_keys", "1292": "vjepa2/modeling_vjepa2.py:VJEPA2RopeAttention", "1293": "vjepa2/modeling_vjepa2.py:drop_path", "1294": "vjepa2/modeling_vjepa2.py:VJEPA2DropPath", "1295": "vjepa2/modeling_vjepa2.py:VJEPA2MLP", "1296": "vjepa2/modeling_vjepa2.py:VJEPA2Layer", "1297": "vjepa2/modeling_vjepa2.py:VJEPA2Encoder", "1298": "vjepa2/modeling_vjepa2.py:apply_masks", "1299": "vjepa2/modeling_vjepa2.py:VJEPA2PredictorEmbeddings", "1300": "vjepa2/modeling_vjepa2.py:VJEPA2Predictor", "1301": "vjepa2/modeling_vjepa2.py:VJEPA2PoolerSelfAttention", "1302": "vjepa2/modeling_vjepa2.py:VJEPA2PoolerCrossAttention", "1303": "vjepa2/modeling_vjepa2.py:VJEPA2PoolerSelfAttentionLayer", "1304": "vjepa2/modeling_vjepa2.py:VJEPA2PoolerCrossAttentionLayer", "1305": "vjepa2/modeling_vjepa2.py:VJEPA2AttentivePooler", "1306": "vjepa2/modeling_vjepa2.py:VJEPA2PreTrainedModel", "1307": "vjepa2/modeling_vjepa2.py:_convert_head_mask_to_5d", "1308": "vjepa2/modeling_vjepa2.py:VJEPA2Model", "1309": "vjepa2/modeling_vjepa2.py:VJEPA2ForVideoClassification", "1310": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1RMSNorm", "1311": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1MLP", "1312": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:rotate_half", "1313": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:apply_rotary_pos_emb", "1314": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:repeat_kv", "1315": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:eager_attention_forward", "1316": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1Attention", "1317": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1Gate", "1318": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1Moe", "1319": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1DecoderLayer", "1320": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1PreTrainedModel", "1321": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1RotaryEmbedding", "1322": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1Model", "1323": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1ForCausalLM", "1324": "hunyuan_v1_moe/modeling_hunyuan_v1_moe.py:HunYuanMoEV1ForSequenceClassification", "1325": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextRMSNorm", "1326": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextRouter", "1327": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextExperts", "1328": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextSparseMoeBlock", "1329": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:rotate_half", "1330": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:repeat_kv", "1331": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:eager_attention_forward", "1332": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:apply_rotary_pos_emb", "1333": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextAttention", "1334": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextMLP", "1335": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextDecoderLayer", "1336": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoePreTrainedModel", "1337": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeVisionMLP", "1338": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeVisionPatchEmbed", "1339": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeVisionRotaryEmbedding", "1340": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeVisionPatchMerger", "1341": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:apply_rotary_pos_emb_vision", "1342": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeVisionAttention", "1343": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeVisionBlock", "1344": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeVisionModel", "1345": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextRotaryEmbedding", "1346": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeTextModel", "1347": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeModelOutputWithPast", "1348": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeModel", "1349": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeCausalLMOutputWithPast", "1350": "qwen3_vl_moe/modeling_qwen3_vl_moe.py:Qwen3VLMoeForConditionalGeneration", "1351": "evolla/modeling_evolla.py:create_position_ids_from_input_ids", "1352": "evolla/modeling_evolla.py:EvollaSaProtEmbeddings", "1353": "evolla/modeling_evolla.py:rotate_half_esm", "1354": "evolla/modeling_evolla.py:apply_rotary_pos_emb_esm", "1355": "evolla/modeling_evolla.py:EvollaSaProtRotaryEmbedding", "1356": "evolla/modeling_evolla.py:eager_attention_forward", "1357": "evolla/modeling_evolla.py:EvollaSaProtSelfAttention", "1358": "evolla/modeling_evolla.py:EvollaSaProtSelfOutput", "1359": "evolla/modeling_evolla.py:EvollaSaProtAttention", "1360": "evolla/modeling_evolla.py:gelu", "1361": "evolla/modeling_evolla.py:EvollaSaProtIntermediate", "1362": "evolla/modeling_evolla.py:EvollaSaProtOutput", "1363": "evolla/modeling_evolla.py:EvollaSaProtLayer", "1364": "evolla/modeling_evolla.py:EvollaSaProtEncoder", "1365": "evolla/modeling_evolla.py:EvollaSaProtPooler", "1366": "evolla/modeling_evolla.py:EvollaSaProtPreTrainedModel", "1367": "evolla/modeling_evolla.py:EvollaSaProtProteinEncoder", "1368": "evolla/modeling_evolla.py:EvollaSequenceCompressorAttention", "1369": "evolla/modeling_evolla.py:EvollaFeedForward", "1370": "evolla/modeling_evolla.py:EvollaSequenceCompressorResampler", "1371": "evolla/modeling_evolla.py:EvollaProteinEncoderModelOutput", "1372": "evolla/modeling_evolla.py:EvollaProteinEncoder", "1373": "evolla/modeling_evolla.py:EvollaSequenceAlignerCrossAttention", "1374": "evolla/modeling_evolla.py:EvollaRMSNorm", "1375": "evolla/modeling_evolla.py:EvollaRotaryEmbedding", "1376": "evolla/modeling_evolla.py:EvollaMLP", "1377": "evolla/modeling_evolla.py:rotate_half", "1378": "evolla/modeling_evolla.py:apply_rotary_pos_emb", "1379": "evolla/modeling_evolla.py:repeat_kv", "1380": "evolla/modeling_evolla.py:EvollaAttention", "1381": "evolla/modeling_evolla.py:EvollaDecoderLayer", "1382": "evolla/modeling_evolla.py:EvollaPreTrainedModel", "1383": "evolla/modeling_evolla.py:EvollaModel", "1384": "evolla/modeling_evolla.py:EvollaForProteinText2Text", "1385": "sam2/modeling_sam2.py:Sam2VisionEncoderOutput", "1386": "sam2/modeling_sam2.py:Sam2ImageSegmentationOutput", "1387": "sam2/modeling_sam2.py:Sam2PatchEmbeddings", "1388": "sam2/modeling_sam2.py:Sam2SinePositionEmbedding", "1389": "sam2/modeling_sam2.py:Sam2VisionNeck", "1390": "sam2/modeling_sam2.py:eager_attention_forward", "1391": "sam2/modeling_sam2.py:do_pool", "1392": "sam2/modeling_sam2.py:Sam2MultiScaleAttention", "1393": "sam2/modeling_sam2.py:Sam2FeedForward", "1394": "sam2/modeling_sam2.py:window_partition", "1395": "sam2/modeling_sam2.py:window_unpartition", "1396": "sam2/modeling_sam2.py:Sam2MultiScaleBlock", "1397": "sam2/modeling_sam2.py:Sam2HieraDetModelOutput", "1398": "sam2/modeling_sam2.py:Sam2PreTrainedModel", "1399": "sam2/modeling_sam2.py:Sam2HieraDetModel", "1400": "sam2/modeling_sam2.py:Sam2VisionModel", "1401": "sam2/modeling_sam2.py:Sam2PositionalEmbedding", "1402": "sam2/modeling_sam2.py:Sam2MaskEmbedding", "1403": "sam2/modeling_sam2.py:Sam2PromptEncoder", "1404": "sam2/modeling_sam2.py:Sam2Attention", "1405": "sam2/modeling_sam2.py:Sam2TwoWayAttentionBlock", "1406": "sam2/modeling_sam2.py:Sam2TwoWayTransformer", "1407": "sam2/modeling_sam2.py:Sam2LayerNorm", "1408": "sam2/modeling_sam2.py:Sam2MaskDecoder", "1409": "sam2/modeling_sam2.py:Sam2Model", "1410": "pixtral/modeling_pixtral.py:position_ids_in_meshgrid", "1411": "pixtral/modeling_pixtral.py:PixtralRotaryEmbedding", "1412": "pixtral/modeling_pixtral.py:rotate_half", "1413": "pixtral/modeling_pixtral.py:apply_rotary_pos_emb", "1414": "pixtral/modeling_pixtral.py:eager_attention_forward", "1415": "pixtral/modeling_pixtral.py:PixtralAttention", "1416": "pixtral/modeling_pixtral.py:PixtralMLP", "1417": "pixtral/modeling_pixtral.py:PixtralRMSNorm", "1418": "pixtral/modeling_pixtral.py:PixtralAttentionLayer", "1419": "pixtral/modeling_pixtral.py:PixtralTransformer", "1420": "pixtral/modeling_pixtral.py:PixtralPreTrainedModel", "1421": "pixtral/modeling_pixtral.py:generate_block_attention_mask", "1422": "pixtral/modeling_pixtral.py:PixtralVisionModel", "1423": "vit_mae/modeling_vit_mae.py:ViTMAEModelOutput", "1424": "vit_mae/modeling_vit_mae.py:ViTMAEDecoderOutput", "1425": "vit_mae/modeling_vit_mae.py:ViTMAEForPreTrainingOutput", "1426": "vit_mae/modeling_vit_mae.py:get_2d_sincos_pos_embed", "1427": "vit_mae/modeling_vit_mae.py:get_2d_sincos_pos_embed_from_grid", "1428": "vit_mae/modeling_vit_mae.py:get_1d_sincos_pos_embed_from_grid", "1429": "vit_mae/modeling_vit_mae.py:ViTMAEEmbeddings", "1430": "vit_mae/modeling_vit_mae.py:ViTMAEPatchEmbeddings", "1431": "vit_mae/modeling_vit_mae.py:eager_attention_forward", "1432": "vit_mae/modeling_vit_mae.py:ViTMAESelfAttention", "1433": "vit_mae/modeling_vit_mae.py:ViTMAESelfOutput", "1434": "vit_mae/modeling_vit_mae.py:ViTMAEAttention", "1435": "vit_mae/modeling_vit_mae.py:ViTMAEIntermediate", "1436": "vit_mae/modeling_vit_mae.py:ViTMAEOutput", "1437": "vit_mae/modeling_vit_mae.py:ViTMAELayer", "1438": "vit_mae/modeling_vit_mae.py:ViTMAEEncoder", "1439": "vit_mae/modeling_vit_mae.py:ViTMAEPreTrainedModel", "1440": "vit_mae/modeling_vit_mae.py:ViTMAEModel", "1441": "vit_mae/modeling_vit_mae.py:ViTMAEDecoder", "1442": "vit_mae/modeling_vit_mae.py:ViTMAEForPreTraining", "1443": "gemma3n/modeling_gemma3n.py:Gemma3nModelOutputWithPast", "1444": "gemma3n/modeling_gemma3n.py:Gemma3nCausalLMOutputWithPast", "1445": "gemma3n/modeling_gemma3n.py:Gemma3nRMSNorm", "1446": "gemma3n/modeling_gemma3n.py:Gemma3nAudioRelativePositionEmbedding", "1447": "gemma3n/modeling_gemma3n.py:Gemma3nAudioAttention", "1448": "gemma3n/modeling_gemma3n.py:Gemma3nAudioCumulativeGroupNorm", "1449": "gemma3n/modeling_gemma3n.py:Gemma3nAudioSSCPConvBlock", "1450": "gemma3n/modeling_gemma3n.py:Gemma3nAudioSubSampleConvProjection", "1451": "gemma3n/modeling_gemma3n.py:Gemma3nAudioConformerAttention", "1452": "gemma3n/modeling_gemma3n.py:Gemma3nAudioConformerFeedForward", "1453": "gemma3n/modeling_gemma3n.py:Gemma3nAudioConformerLightConv1d", "1454": "gemma3n/modeling_gemma3n.py:Gemma3nAudioConformerBlock", "1455": "gemma3n/modeling_gemma3n.py:Gemma3nAudioEncoder", "1456": "gemma3n/modeling_gemma3n.py:Gemma3nTextScaledWordEmbedding", "1457": "gemma3n/modeling_gemma3n.py:Gemma3nTextLaurelBlock", "1458": "gemma3n/modeling_gemma3n.py:Gemma3nTextMLP", "1459": "gemma3n/modeling_gemma3n.py:Gemma3nTextAltUp", "1460": "gemma3n/modeling_gemma3n.py:Gemma3nTextRotaryEmbedding", "1461": "gemma3n/modeling_gemma3n.py:rotate_half", "1462": "gemma3n/modeling_gemma3n.py:repeat_kv", "1463": "gemma3n/modeling_gemma3n.py:eager_attention_forward", "1464": "gemma3n/modeling_gemma3n.py:apply_rotary_pos_emb", "1465": "gemma3n/modeling_gemma3n.py:Gemma3nTextAttention", "1466": "gemma3n/modeling_gemma3n.py:Gemma3nTextDecoderLayer", "1467": "gemma3n/modeling_gemma3n.py:Gemma3nPreTrainedModel", "1468": "gemma3n/modeling_gemma3n.py:Gemma3nTextModel", "1469": "gemma3n/modeling_gemma3n.py:Gemma3nForCausalLM", "1470": "gemma3n/modeling_gemma3n.py:Gemma3nMultimodalEmbedder", "1471": "gemma3n/modeling_gemma3n.py:Gemma3nModel", "1472": "gemma3n/modeling_gemma3n.py:Gemma3nForConditionalGeneration", "1473": "persimmon/modeling_persimmon.py:PersimmonRotaryEmbedding", "1474": "persimmon/modeling_persimmon.py:rotate_half", "1475": "persimmon/modeling_persimmon.py:apply_rotary_pos_emb", "1476": "persimmon/modeling_persimmon.py:PersimmonMLP", "1477": "persimmon/modeling_persimmon.py:eager_attention_forward", "1478": "persimmon/modeling_persimmon.py:PersimmonAttention", "1479": "persimmon/modeling_persimmon.py:PersimmonDecoderLayer", "1480": "persimmon/modeling_persimmon.py:PersimmonPreTrainedModel", "1481": "persimmon/modeling_persimmon.py:PersimmonModel", "1482": "persimmon/modeling_persimmon.py:PersimmonForCausalLM", "1483": "persimmon/modeling_persimmon.py:PersimmonForSequenceClassification", "1484": "persimmon/modeling_persimmon.py:PersimmonForTokenClassification", "1485": "xlm/modeling_xlm.py:create_sinusoidal_embeddings", "1486": "xlm/modeling_xlm.py:get_masks", "1487": "xlm/modeling_xlm.py:XLMSquadHeadOutput", "1488": "xlm/modeling_xlm.py:XLMPoolerStartLogits", "1489": "xlm/modeling_xlm.py:XLMPoolerEndLogits", "1490": "xlm/modeling_xlm.py:XLMPoolerAnswerClass", "1491": "xlm/modeling_xlm.py:XLMSQuADHead", "1492": "xlm/modeling_xlm.py:XLMSequenceSummary", "1493": "xlm/modeling_xlm.py:MultiHeadAttention", "1494": "xlm/modeling_xlm.py:TransformerFFN", "1495": "xlm/modeling_xlm.py:XLMPreTrainedModel", "1496": "xlm/modeling_xlm.py:XLMForQuestionAnsweringOutput", "1497": "xlm/modeling_xlm.py:XLMModel", "1498": "xlm/modeling_xlm.py:XLMPredLayer", "1499": "xlm/modeling_xlm.py:XLMWithLMHeadModel", "1500": "xlm/modeling_xlm.py:XLMForSequenceClassification", "1501": "xlm/modeling_xlm.py:XLMForQuestionAnsweringSimple", "1502": "xlm/modeling_xlm.py:XLMForQuestionAnswering", "1503": "xlm/modeling_xlm.py:XLMForTokenClassification", "1504": "xlm/modeling_xlm.py:XLMForMultipleChoice", "1505": "xmod/modeling_xmod.py:XmodEmbeddings", "1506": "xmod/modeling_xmod.py:eager_attention_forward", "1507": "xmod/modeling_xmod.py:XmodSelfAttention", "1508": "xmod/modeling_xmod.py:XmodCrossAttention", "1509": "xmod/modeling_xmod.py:XmodSelfOutput", "1510": "xmod/modeling_xmod.py:XmodAttention", "1511": "xmod/modeling_xmod.py:XmodIntermediate", "1512": "xmod/modeling_xmod.py:XmodAdapter", "1513": "xmod/modeling_xmod.py:XmodOutput", "1514": "xmod/modeling_xmod.py:XmodLayer", "1515": "xmod/modeling_xmod.py:XmodEncoder", "1516": "xmod/modeling_xmod.py:XmodPooler", "1517": "xmod/modeling_xmod.py:XmodPreTrainedModel", "1518": "xmod/modeling_xmod.py:XmodModel", "1519": "xmod/modeling_xmod.py:XmodForCausalLM", "1520": "xmod/modeling_xmod.py:XmodForMaskedLM", "1521": "xmod/modeling_xmod.py:XmodLMHead", "1522": "xmod/modeling_xmod.py:XmodForSequenceClassification", "1523": "xmod/modeling_xmod.py:XmodForMultipleChoice", "1524": "xmod/modeling_xmod.py:XmodForTokenClassification", "1525": "xmod/modeling_xmod.py:XmodClassificationHead", "1526": "xmod/modeling_xmod.py:XmodForQuestionAnswering", "1527": "roberta/modeling_roberta.py:RobertaEmbeddings", "1528": "roberta/modeling_roberta.py:eager_attention_forward", "1529": "roberta/modeling_roberta.py:RobertaSelfAttention", "1530": "roberta/modeling_roberta.py:RobertaCrossAttention", "1531": "roberta/modeling_roberta.py:RobertaSelfOutput", "1532": "roberta/modeling_roberta.py:RobertaAttention", "1533": "roberta/modeling_roberta.py:RobertaIntermediate", "1534": "roberta/modeling_roberta.py:RobertaOutput", "1535": "roberta/modeling_roberta.py:RobertaLayer", "1536": "roberta/modeling_roberta.py:RobertaPreTrainedModel", "1537": "roberta/modeling_roberta.py:RobertaEncoder", "1538": "roberta/modeling_roberta.py:RobertaPooler", "1539": "roberta/modeling_roberta.py:RobertaModel", "1540": "roberta/modeling_roberta.py:RobertaForCausalLM", "1541": "roberta/modeling_roberta.py:RobertaForMaskedLM", "1542": "roberta/modeling_roberta.py:RobertaLMHead", "1543": "roberta/modeling_roberta.py:RobertaForSequenceClassification", "1544": "roberta/modeling_roberta.py:RobertaForMultipleChoice", "1545": "roberta/modeling_roberta.py:RobertaForTokenClassification", "1546": "roberta/modeling_roberta.py:RobertaClassificationHead", "1547": "roberta/modeling_roberta.py:RobertaForQuestionAnswering", "1548": "csm/modeling_csm.py:CsmOutputWithPast", "1549": "csm/modeling_csm.py:CsmRMSNorm", "1550": "csm/modeling_csm.py:CsmRotaryEmbedding", "1551": "csm/modeling_csm.py:CsmMLP", "1552": "csm/modeling_csm.py:rotate_half", "1553": "csm/modeling_csm.py:apply_rotary_pos_emb", "1554": "csm/modeling_csm.py:repeat_kv", "1555": "csm/modeling_csm.py:eager_attention_forward", "1556": "csm/modeling_csm.py:CsmAttention", "1557": "csm/modeling_csm.py:CsmDecoderLayer", "1558": "csm/modeling_csm.py:CsmPreTrainedModel", "1559": "csm/modeling_csm.py:CsmDepthDecoderModel", "1560": "csm/modeling_csm.py:CsmCodebooksHead", "1561": "csm/modeling_csm.py:CsmDepthDecoderForCausalLM", "1562": "csm/modeling_csm.py:CsmBackboneModelEmbeddings", "1563": "csm/modeling_csm.py:CsmBackboneModel", "1564": "csm/modeling_csm.py:CsmForConditionalGeneration", "1565": "mra/modeling_mra.py:load_cuda_kernels", "1566": "mra/modeling_mra.py:sparse_max", "1567": "mra/modeling_mra.py:sparse_mask", "1568": "mra/modeling_mra.py:mm_to_sparse", "1569": "mra/modeling_mra.py:sparse_dense_mm", "1570": "mra/modeling_mra.py:transpose_indices", "1571": "mra/modeling_mra.py:MraSampledDenseMatMul", "1572": "mra/modeling_mra.py:MraSparseDenseMatMul", "1573": "mra/modeling_mra.py:MraReduceSum", "1574": "mra/modeling_mra.py:get_low_resolution_logit", "1575": "mra/modeling_mra.py:get_block_idxes", "1576": "mra/modeling_mra.py:mra2_attention", "1577": "mra/modeling_mra.py:MraEmbeddings", "1578": "mra/modeling_mra.py:MraSelfAttention", "1579": "mra/modeling_mra.py:MraSelfOutput", "1580": "mra/modeling_mra.py:MraAttention", "1581": "mra/modeling_mra.py:MraIntermediate", "1582": "mra/modeling_mra.py:MraOutput", "1583": "mra/modeling_mra.py:MraLayer", "1584": "mra/modeling_mra.py:MraEncoder", "1585": "mra/modeling_mra.py:MraPredictionHeadTransform", "1586": "mra/modeling_mra.py:MraLMPredictionHead", "1587": "mra/modeling_mra.py:MraOnlyMLMHead", "1588": "mra/modeling_mra.py:MraPreTrainedModel", "1589": "mra/modeling_mra.py:MraModel", "1590": "mra/modeling_mra.py:MraForMaskedLM", "1591": "mra/modeling_mra.py:MraClassificationHead", "1592": "mra/modeling_mra.py:MraForSequenceClassification", "1593": "mra/modeling_mra.py:MraForMultipleChoice", "1594": "mra/modeling_mra.py:MraForTokenClassification", "1595": "mra/modeling_mra.py:MraForQuestionAnswering", "1596": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTEmbeddings", "1597": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTPatchEmbeddings", "1598": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:eager_attention_forward", "1599": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTSelfAttention", "1600": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTSelfOutput", "1601": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTAttention", "1602": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTIntermediate", "1603": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTOutput", "1604": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTLayer", "1605": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTEncoder", "1606": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTPreTrainedModel", "1607": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTModel", "1608": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTMLPHead", "1609": "audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py:ASTForAudioClassification", "1610": "owlv2/modeling_owlv2.py:contrastive_loss", "1611": "owlv2/modeling_owlv2.py:owlv2_loss", "1612": "owlv2/modeling_owlv2.py:Owlv2Output", "1613": "owlv2/modeling_owlv2.py:_upcast", "1614": "owlv2/modeling_owlv2.py:box_area", "1615": "owlv2/modeling_owlv2.py:box_iou", "1616": "owlv2/modeling_owlv2.py:generalized_box_iou", "1617": "owlv2/modeling_owlv2.py:Owlv2ObjectDetectionOutput", "1618": "owlv2/modeling_owlv2.py:Owlv2ImageGuidedObjectDetectionOutput", "1619": "owlv2/modeling_owlv2.py:Owlv2VisionEmbeddings", "1620": "owlv2/modeling_owlv2.py:Owlv2TextEmbeddings", "1621": "owlv2/modeling_owlv2.py:Owlv2Attention", "1622": "owlv2/modeling_owlv2.py:Owlv2MLP", "1623": "owlv2/modeling_owlv2.py:Owlv2EncoderLayer", "1624": "owlv2/modeling_owlv2.py:Owlv2PreTrainedModel", "1625": "owlv2/modeling_owlv2.py:Owlv2Encoder", "1626": "owlv2/modeling_owlv2.py:Owlv2TextTransformer", "1627": "owlv2/modeling_owlv2.py:Owlv2TextModel", "1628": "owlv2/modeling_owlv2.py:Owlv2VisionTransformer", "1629": "owlv2/modeling_owlv2.py:Owlv2VisionModel", "1630": "owlv2/modeling_owlv2.py:Owlv2Model", "1631": "owlv2/modeling_owlv2.py:Owlv2BoxPredictionHead", "1632": "owlv2/modeling_owlv2.py:Owlv2ClassPredictionHead", "1633": "owlv2/modeling_owlv2.py:Owlv2ForObjectDetection", "1634": "decision_transformer/modeling_decision_transformer.py:eager_attention_forward", "1635": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerGPT2Attention", "1636": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerGPT2MLP", "1637": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerGPT2Block", "1638": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerGPT2PreTrainedModel", "1639": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerGPT2Model", "1640": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerOutput", "1641": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerPreTrainedModel", "1642": "decision_transformer/modeling_decision_transformer.py:DecisionTransformerModel", "1643": "mpt/modeling_mpt.py:build_mpt_alibi_tensor", "1644": "mpt/modeling_mpt.py:MptAttention", "1645": "mpt/modeling_mpt.py:MptMLP", "1646": "mpt/modeling_mpt.py:MptBlock", "1647": "mpt/modeling_mpt.py:MptPreTrainedModel", "1648": "mpt/modeling_mpt.py:MptModel", "1649": "mpt/modeling_mpt.py:MptForCausalLM", "1650": "mpt/modeling_mpt.py:MptForSequenceClassification", "1651": "mpt/modeling_mpt.py:MptForTokenClassification", "1652": "mpt/modeling_mpt.py:MptForQuestionAnswering", "1653": "clip/modeling_clip.py:contrastive_loss", "1654": "clip/modeling_clip.py:clip_loss", "1655": "clip/modeling_clip.py:_get_vector_norm", "1656": "clip/modeling_clip.py:CLIPVisionModelOutput", "1657": "clip/modeling_clip.py:CLIPTextModelOutput", "1658": "clip/modeling_clip.py:CLIPOutput", "1659": "clip/modeling_clip.py:CLIPVisionEmbeddings", "1660": "clip/modeling_clip.py:CLIPTextEmbeddings", "1661": "clip/modeling_clip.py:eager_attention_forward", "1662": "clip/modeling_clip.py:CLIPAttention", "1663": "clip/modeling_clip.py:CLIPMLP", "1664": "clip/modeling_clip.py:CLIPEncoderLayer", "1665": "clip/modeling_clip.py:CLIPPreTrainedModel", "1666": "clip/modeling_clip.py:CLIPEncoder", "1667": "clip/modeling_clip.py:CLIPTextTransformer", "1668": "clip/modeling_clip.py:CLIPTextModel", "1669": "clip/modeling_clip.py:CLIPVisionTransformer", "1670": "clip/modeling_clip.py:CLIPVisionModel", "1671": "clip/modeling_clip.py:CLIPModel", "1672": "clip/modeling_clip.py:CLIPTextModelWithProjection", "1673": "clip/modeling_clip.py:CLIPVisionModelWithProjection", "1674": "clip/modeling_clip.py:CLIPForImageClassification", "1675": "zamba2/modeling_zamba2.py:Zamba2RMSNormGated", "1676": "zamba2/modeling_zamba2.py:Zamba2RMSNorm", "1677": "zamba2/modeling_zamba2.py:Zamba2HybridDynamicCache", "1678": "zamba2/modeling_zamba2.py:Zamba2RotaryEmbedding", "1679": "zamba2/modeling_zamba2.py:repeat_kv", "1680": "zamba2/modeling_zamba2.py:eager_attention_forward", "1681": "zamba2/modeling_zamba2.py:rotate_half", "1682": "zamba2/modeling_zamba2.py:apply_rotary_pos_emb", "1683": "zamba2/modeling_zamba2.py:Zamba2Attention", "1684": "zamba2/modeling_zamba2.py:pad_tensor_by_size", "1685": "zamba2/modeling_zamba2.py:reshape_into_chunks", "1686": "zamba2/modeling_zamba2.py:segment_sum", "1687": "zamba2/modeling_zamba2.py:Zamba2MambaMixer", "1688": "zamba2/modeling_zamba2.py:Zamba2MLP", "1689": "zamba2/modeling_zamba2.py:Zamba2AttentionDecoderLayer", "1690": "zamba2/modeling_zamba2.py:Zamba2MambaDecoderLayer", "1691": "zamba2/modeling_zamba2.py:Zamba2HybridLayer", "1692": "zamba2/modeling_zamba2.py:Zamba2PreTrainedModel", "1693": "zamba2/modeling_zamba2.py:Zamba2Model", "1694": "zamba2/modeling_zamba2.py:Zamba2ForCausalLM", "1695": "zamba2/modeling_zamba2.py:Zamba2ForSequenceClassification", "1696": "janus/modeling_janus.py:JanusPreTrainedModel", "1697": "janus/modeling_janus.py:JanusVQVAEOutput", "1698": "janus/modeling_janus.py:JanusBaseModelOutputWithPast", "1699": "janus/modeling_janus.py:JanusCausalLMOutputWithPast", "1700": "janus/modeling_janus.py:JanusVisionEmbeddings", "1701": "janus/modeling_janus.py:repeat_kv", "1702": "janus/modeling_janus.py:eager_attention_forward", "1703": "janus/modeling_janus.py:JanusVisionAttention", "1704": "janus/modeling_janus.py:JanusVisionMLP", "1705": "janus/modeling_janus.py:JanusVisionEncoderLayer", "1706": "janus/modeling_janus.py:JanusVisionEncoder", "1707": "janus/modeling_janus.py:JanusAttention", "1708": "janus/modeling_janus.py:JanusMLP", "1709": "janus/modeling_janus.py:JanusEncoderLayer", "1710": "janus/modeling_janus.py:JanusVisionModel", "1711": "janus/modeling_janus.py:JanusVisionAlignerMLP", "1712": "janus/modeling_janus.py:JanusVQVAEVectorQuantizer", "1713": "janus/modeling_janus.py:JanusVQVAEResnetBlock", "1714": "janus/modeling_janus.py:JanusVQVAEAttnBlock", "1715": "janus/modeling_janus.py:JanusVQVAEConvDownsample", "1716": "janus/modeling_janus.py:JanusVQVAEConvUpsample", "1717": "janus/modeling_janus.py:JanusVQVAEMidBlock", "1718": "janus/modeling_janus.py:JanusVQVAEEncoder", "1719": "janus/modeling_janus.py:JanusVQVAEDecoder", "1720": "janus/modeling_janus.py:JanusVQVAE", "1721": "janus/modeling_janus.py:JanusVQVAEAlignerMLP", "1722": "janus/modeling_janus.py:JanusVQVAEHead", "1723": "janus/modeling_janus.py:JanusModel", "1724": "janus/modeling_janus.py:JanusForConditionalGeneration", "1725": "gpt_bigcode/modeling_gpt_bigcode.py:upcast_masked_softmax", "1726": "gpt_bigcode/modeling_gpt_bigcode.py:upcast_softmax", "1727": "gpt_bigcode/modeling_gpt_bigcode.py:masked_softmax", "1728": "gpt_bigcode/modeling_gpt_bigcode.py:repeat_kv", "1729": "gpt_bigcode/modeling_gpt_bigcode.py:eager_attention_forward", "1730": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodeAttention", "1731": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodeMLP", "1732": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodeBlock", "1733": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodePreTrainedModel", "1734": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodeModel", "1735": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodeForCausalLM", "1736": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodeForSequenceClassification", "1737": "gpt_bigcode/modeling_gpt_bigcode.py:GPTBigCodeForTokenClassification", "1738": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerForPreTrainingOutput", "1739": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerSamePadLayer", "1740": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerPositionalConvEmbedding", "1741": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerRotaryPositionalEmbedding", "1742": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerRelPositionalEmbedding", "1743": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerNoLayerNormConvLayer", "1744": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerLayerNormConvLayer", "1745": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerGroupNormConvLayer", "1746": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerFeatureEncoder", "1747": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerFeatureProjection", "1748": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerFeedForward", "1749": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerConvolutionModule", "1750": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerSelfAttention", "1751": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerEncoderLayer", "1752": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerEncoder", "1753": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerGumbelVectorQuantizer", "1754": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerAdapter", "1755": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerAdapterLayer", "1756": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerPreTrainedModel", "1757": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:_compute_mask_indices", "1758": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerModel", "1759": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerForPreTraining", "1760": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerForCTC", "1761": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerForSequenceClassification", "1762": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerForAudioFrameClassification", "1763": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:AMSoftmaxLoss", "1764": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:TDNNLayer", "1765": "wav2vec2_conformer/modeling_wav2vec2_conformer.py:Wav2Vec2ConformerForXVector", "1766": "mlcd/modeling_mlcd.py:MLCDMLP", "1767": "mlcd/modeling_mlcd.py:MLCDRotaryEmbedding", "1768": "mlcd/modeling_mlcd.py:MLCDVisionEmbeddings", "1769": "mlcd/modeling_mlcd.py:eager_attention_forward", "1770": "mlcd/modeling_mlcd.py:rotate_half", "1771": "mlcd/modeling_mlcd.py:repeat_kv", "1772": "mlcd/modeling_mlcd.py:apply_rotary_pos_emb_vision", "1773": "mlcd/modeling_mlcd.py:MLCDAttention", "1774": "mlcd/modeling_mlcd.py:MLCDEncoderLayer", "1775": "mlcd/modeling_mlcd.py:MLCDEncoder", "1776": "mlcd/modeling_mlcd.py:MLCDVisionTransformer", "1777": "mlcd/modeling_mlcd.py:MLCDPreTrainedModel", "1778": "mlcd/modeling_mlcd.py:MLCDVisionModel", "1779": "vits/modeling_vits.py:VitsModelOutput", "1780": "vits/modeling_vits.py:VitsTextEncoderOutput", "1781": "vits/modeling_vits.py:fused_add_tanh_sigmoid_multiply", "1782": "vits/modeling_vits.py:_unconstrained_rational_quadratic_spline", "1783": "vits/modeling_vits.py:_rational_quadratic_spline", "1784": "vits/modeling_vits.py:VitsWaveNet", "1785": "vits/modeling_vits.py:VitsPosteriorEncoder", "1786": "vits/modeling_vits.py:HifiGanResidualBlock", "1787": "vits/modeling_vits.py:VitsHifiGan", "1788": "vits/modeling_vits.py:VitsResidualCouplingLayer", "1789": "vits/modeling_vits.py:VitsResidualCouplingBlock", "1790": "vits/modeling_vits.py:VitsDilatedDepthSeparableConv", "1791": "vits/modeling_vits.py:VitsConvFlow", "1792": "vits/modeling_vits.py:VitsElementwiseAffine", "1793": "vits/modeling_vits.py:VitsStochasticDurationPredictor", "1794": "vits/modeling_vits.py:VitsDurationPredictor", "1795": "vits/modeling_vits.py:VitsAttention", "1796": "vits/modeling_vits.py:VitsFeedForward", "1797": "vits/modeling_vits.py:VitsEncoderLayer", "1798": "vits/modeling_vits.py:VitsEncoder", "1799": "vits/modeling_vits.py:VitsTextEncoder", "1800": "vits/modeling_vits.py:VitsPreTrainedModel", "1801": "vits/modeling_vits.py:VitsModel", "1802": "encodec/modeling_encodec.py:EncodecOutput", "1803": "encodec/modeling_encodec.py:EncodecEncoderOutput", "1804": "encodec/modeling_encodec.py:EncodecDecoderOutput", "1805": "encodec/modeling_encodec.py:EncodecConv1d", "1806": "encodec/modeling_encodec.py:EncodecConvTranspose1d", "1807": "encodec/modeling_encodec.py:EncodecLSTM", "1808": "encodec/modeling_encodec.py:EncodecResnetBlock", "1809": "encodec/modeling_encodec.py:EncodecEncoder", "1810": "encodec/modeling_encodec.py:EncodecDecoder", "1811": "encodec/modeling_encodec.py:EncodecEuclideanCodebook", "1812": "encodec/modeling_encodec.py:EncodecVectorQuantization", "1813": "encodec/modeling_encodec.py:EncodecResidualVectorQuantizer", "1814": "encodec/modeling_encodec.py:EncodecPreTrainedModel", "1815": "encodec/modeling_encodec.py:EncodecModel", "1816": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLEmbeddings", "1817": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:eager_attention_forward", "1818": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLSelfAttention", "1819": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLCrossAttention", "1820": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLSelfOutput", "1821": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLAttention", "1822": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLOutput", "1823": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLIntermediate", "1824": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLLayer", "1825": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLEncoder", "1826": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLPreTrainedModel", "1827": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLPooler", "1828": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLModel", "1829": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLLMHead", "1830": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLClassificationHead", "1831": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLForCausalLM", "1832": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLForMaskedLM", "1833": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLForSequenceClassification", "1834": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLForMultipleChoice", "1835": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLForTokenClassification", "1836": "xlm_roberta_xl/modeling_xlm_roberta_xl.py:XLMRobertaXLForQuestionAnswering", "1837": "gemma3/modeling_gemma3.py:Gemma3ModelOutputWithPast", "1838": "gemma3/modeling_gemma3.py:Gemma3CausalLMOutputWithPast", "1839": "gemma3/modeling_gemma3.py:Gemma3TextScaledWordEmbedding", "1840": "gemma3/modeling_gemma3.py:Gemma3MLP", "1841": "gemma3/modeling_gemma3.py:Gemma3RMSNorm", "1842": "gemma3/modeling_gemma3.py:Gemma3RotaryEmbedding", "1843": "gemma3/modeling_gemma3.py:rotate_half", "1844": "gemma3/modeling_gemma3.py:apply_rotary_pos_emb", "1845": "gemma3/modeling_gemma3.py:repeat_kv", "1846": "gemma3/modeling_gemma3.py:eager_attention_forward", "1847": "gemma3/modeling_gemma3.py:Gemma3Attention", "1848": "gemma3/modeling_gemma3.py:Gemma3DecoderLayer", "1849": "gemma3/modeling_gemma3.py:Gemma3PreTrainedModel", "1850": "gemma3/modeling_gemma3.py:_bidirectional_window_overlay", "1851": "gemma3/modeling_gemma3.py:Gemma3TextModel", "1852": "gemma3/modeling_gemma3.py:Gemma3ForCausalLM", "1853": "gemma3/modeling_gemma3.py:Gemma3MultiModalProjector", "1854": "gemma3/modeling_gemma3.py:token_type_ids_mask_function", "1855": "gemma3/modeling_gemma3.py:create_causal_mask_mapping", "1856": "gemma3/modeling_gemma3.py:Gemma3Model", "1857": "gemma3/modeling_gemma3.py:Gemma3ForConditionalGeneration", "1858": "gemma3/modeling_gemma3.py:Gemma3ForSequenceClassification", "1859": "gemma3/modeling_gemma3.py:Gemma3TextForSequenceClassification", "1860": "big_bird/modeling_big_bird.py:BigBirdEmbeddings", "1861": "big_bird/modeling_big_bird.py:BigBirdSelfAttention", "1862": "big_bird/modeling_big_bird.py:BigBirdBlockSparseAttention", "1863": "big_bird/modeling_big_bird.py:BigBirdSelfOutput", "1864": "big_bird/modeling_big_bird.py:BigBirdAttention", "1865": "big_bird/modeling_big_bird.py:BigBirdIntermediate", "1866": "big_bird/modeling_big_bird.py:BigBirdOutput", "1867": "big_bird/modeling_big_bird.py:BigBirdLayer", "1868": "big_bird/modeling_big_bird.py:BigBirdEncoder", "1869": "big_bird/modeling_big_bird.py:BigBirdPredictionHeadTransform", "1870": "big_bird/modeling_big_bird.py:BigBirdLMPredictionHead", "1871": "big_bird/modeling_big_bird.py:BigBirdOnlyMLMHead", "1872": "big_bird/modeling_big_bird.py:BigBirdOnlyNSPHead", "1873": "big_bird/modeling_big_bird.py:BigBirdPreTrainingHeads", "1874": "big_bird/modeling_big_bird.py:BigBirdPreTrainedModel", "1875": "big_bird/modeling_big_bird.py:BigBirdForPreTrainingOutput", "1876": "big_bird/modeling_big_bird.py:BigBirdForQuestionAnsweringModelOutput", "1877": "big_bird/modeling_big_bird.py:BigBirdModel", "1878": "big_bird/modeling_big_bird.py:BigBirdForPreTraining", "1879": "big_bird/modeling_big_bird.py:BigBirdForMaskedLM", "1880": "big_bird/modeling_big_bird.py:BigBirdForCausalLM", "1881": "big_bird/modeling_big_bird.py:BigBirdClassificationHead", "1882": "big_bird/modeling_big_bird.py:BigBirdForSequenceClassification", "1883": "big_bird/modeling_big_bird.py:BigBirdForMultipleChoice", "1884": "big_bird/modeling_big_bird.py:BigBirdForTokenClassification", "1885": "big_bird/modeling_big_bird.py:BigBirdForQuestionAnsweringHead", "1886": "big_bird/modeling_big_bird.py:BigBirdForQuestionAnswering", "1887": "ovis2/modeling_ovis2.py:Ovis2ModelOutputWithPast", "1888": "ovis2/modeling_ovis2.py:Ovis2CausalLMOutputWithPast", "1889": "ovis2/modeling_ovis2.py:Ovis2RMSNorm", "1890": "ovis2/modeling_ovis2.py:Ovis2VisionMLP", "1891": "ovis2/modeling_ovis2.py:Ovis2VisionEmbeddings", "1892": "ovis2/modeling_ovis2.py:eager_attention_forward", "1893": "ovis2/modeling_ovis2.py:Ovis2VisionAttention", "1894": "ovis2/modeling_ovis2.py:Ovis2MLP", "1895": "ovis2/modeling_ovis2.py:Ovis2Attention", "1896": "ovis2/modeling_ovis2.py:Ovis2VisionEncoderLayer", "1897": "ovis2/modeling_ovis2.py:Ovis2VisionEncoder", "1898": "ovis2/modeling_ovis2.py:Ovis2VisionTransformer", "1899": "ovis2/modeling_ovis2.py:Ovis2VisualEmbeddingTable", "1900": "ovis2/modeling_ovis2.py:Ovis2PreTrainedModel", "1901": "ovis2/modeling_ovis2.py:hard_softmax", "1902": "ovis2/modeling_ovis2.py:Ovis2VisionModel", "1903": "ovis2/modeling_ovis2.py:Ovis2Model", "1904": "ovis2/modeling_ovis2.py:Ovis2ForConditionalGeneration", "1905": "convnextv2/modeling_convnextv2.py:drop_path", "1906": "convnextv2/modeling_convnextv2.py:ConvNextV2DropPath", "1907": "convnextv2/modeling_convnextv2.py:ConvNextV2GRN", "1908": "convnextv2/modeling_convnextv2.py:ConvNextV2LayerNorm", "1909": "convnextv2/modeling_convnextv2.py:ConvNextV2Embeddings", "1910": "convnextv2/modeling_convnextv2.py:ConvNextV2Layer", "1911": "convnextv2/modeling_convnextv2.py:ConvNextV2Stage", "1912": "convnextv2/modeling_convnextv2.py:ConvNextV2Encoder", "1913": "convnextv2/modeling_convnextv2.py:ConvNextV2PreTrainedModel", "1914": "convnextv2/modeling_convnextv2.py:ConvNextV2Model", "1915": "convnextv2/modeling_convnextv2.py:ConvNextV2ForImageClassification", "1916": "convnextv2/modeling_convnextv2.py:ConvNextV2Backbone", "1917": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoVisionEmbeddings", "1918": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoPreTrainedModel", "1919": "instructblipvideo/modeling_instructblipvideo.py:eager_attention_forward", "1920": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoAttention", "1921": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoMLP", "1922": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoEncoderLayer", "1923": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoEncoder", "1924": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoVisionModel", "1925": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerMultiHeadAttention", "1926": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerSelfOutput", "1927": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerAttention", "1928": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerIntermediate", "1929": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerOutput", "1930": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerLayer", "1931": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerEncoder", "1932": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerEmbeddings", "1933": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoQFormerModel", "1934": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoForConditionalGenerationModelOutput", "1935": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoModel", "1936": "instructblipvideo/modeling_instructblipvideo.py:InstructBlipVideoForConditionalGeneration", "1937": "megatron_bert/modeling_megatron_bert.py:MegatronBertEmbeddings", "1938": "megatron_bert/modeling_megatron_bert.py:MegatronBertSelfAttention", "1939": "megatron_bert/modeling_megatron_bert.py:MegatronBertSelfOutput", "1940": "megatron_bert/modeling_megatron_bert.py:MegatronBertAttention", "1941": "megatron_bert/modeling_megatron_bert.py:MegatronBertIntermediate", "1942": "megatron_bert/modeling_megatron_bert.py:MegatronBertOutput", "1943": "megatron_bert/modeling_megatron_bert.py:MegatronBertLayer", "1944": "megatron_bert/modeling_megatron_bert.py:MegatronBertEncoder", "1945": "megatron_bert/modeling_megatron_bert.py:MegatronBertPooler", "1946": "megatron_bert/modeling_megatron_bert.py:MegatronBertPredictionHeadTransform", "1947": "megatron_bert/modeling_megatron_bert.py:MegatronBertLMPredictionHead", "1948": "megatron_bert/modeling_megatron_bert.py:MegatronBertOnlyMLMHead", "1949": "megatron_bert/modeling_megatron_bert.py:MegatronBertOnlyNSPHead", "1950": "megatron_bert/modeling_megatron_bert.py:MegatronBertPreTrainingHeads", "1951": "megatron_bert/modeling_megatron_bert.py:MegatronBertPreTrainedModel", "1952": "megatron_bert/modeling_megatron_bert.py:MegatronBertForPreTrainingOutput", "1953": "megatron_bert/modeling_megatron_bert.py:MegatronBertModel", "1954": "megatron_bert/modeling_megatron_bert.py:MegatronBertForPreTraining", "1955": "megatron_bert/modeling_megatron_bert.py:MegatronBertForCausalLM", "1956": "megatron_bert/modeling_megatron_bert.py:MegatronBertForMaskedLM", "1957": "megatron_bert/modeling_megatron_bert.py:MegatronBertForNextSentencePrediction", "1958": "megatron_bert/modeling_megatron_bert.py:MegatronBertForSequenceClassification", "1959": "megatron_bert/modeling_megatron_bert.py:MegatronBertForMultipleChoice", "1960": "megatron_bert/modeling_megatron_bert.py:MegatronBertForTokenClassification", "1961": "megatron_bert/modeling_megatron_bert.py:MegatronBertForQuestionAnswering", "1962": "longcat_flash/modeling_longcat_flash.py:LongcatFlashRMSNorm", "1963": "longcat_flash/modeling_longcat_flash.py:LongcatFlashRotaryEmbedding", "1964": "longcat_flash/modeling_longcat_flash.py:LongcatFlashMLP", "1965": "longcat_flash/modeling_longcat_flash.py:LongcatFlashTopkRouter", "1966": "longcat_flash/modeling_longcat_flash.py:LongcatFlashMoE", "1967": "longcat_flash/modeling_longcat_flash.py:rotate_half", "1968": "longcat_flash/modeling_longcat_flash.py:repeat_kv", "1969": "longcat_flash/modeling_longcat_flash.py:eager_attention_forward", "1970": "longcat_flash/modeling_longcat_flash.py:apply_rotary_pos_emb_interleave", "1971": "longcat_flash/modeling_longcat_flash.py:yarn_get_mscale", "1972": "longcat_flash/modeling_longcat_flash.py:LongcatFlashMLA", "1973": "longcat_flash/modeling_longcat_flash.py:LongcatFlashDecoderLayer", "1974": "longcat_flash/modeling_longcat_flash.py:LongcatFlashPreTrainedModel", "1975": "longcat_flash/modeling_longcat_flash.py:LongcatFlashModel", "1976": "longcat_flash/modeling_longcat_flash.py:LongcatFlashForCausalLM", "1977": "clap/modeling_clap.py:interpolate", "1978": "clap/modeling_clap.py:window_partition", "1979": "clap/modeling_clap.py:window_reverse", "1980": "clap/modeling_clap.py:contrastive_loss", "1981": "clap/modeling_clap.py:ClapTextModelOutput", "1982": "clap/modeling_clap.py:ClapAudioModelOutput", "1983": "clap/modeling_clap.py:ClapOutput", "1984": "clap/modeling_clap.py:ClapDropPath", "1985": "clap/modeling_clap.py:ClapAudioAFFBlock", "1986": "clap/modeling_clap.py:ClapAudioPatchEmbed", "1987": "clap/modeling_clap.py:ClapAudioSelfAttention", "1988": "clap/modeling_clap.py:ClapAudioSelfOutput", "1989": "clap/modeling_clap.py:ClapAudioAttention", "1990": "clap/modeling_clap.py:ClapAudioIntermediate", "1991": "clap/modeling_clap.py:ClapAudioOutput", "1992": "clap/modeling_clap.py:ClapAudioLayer", "1993": "clap/modeling_clap.py:ClapAudioStage", "1994": "clap/modeling_clap.py:ClapAudioPatchMerging", "1995": "clap/modeling_clap.py:ClapAudioEncoder", "1996": "clap/modeling_clap.py:ClapProjectionLayer", "1997": "clap/modeling_clap.py:ClapTextEmbeddings", "1998": "clap/modeling_clap.py:eager_attention_forward", "1999": "clap/modeling_clap.py:ClapTextSelfAttention", "2000": "clap/modeling_clap.py:ClapTextSelfOutput", "2001": "clap/modeling_clap.py:ClapTextAttention", "2002": "clap/modeling_clap.py:ClapTextIntermediate", "2003": "clap/modeling_clap.py:ClapTextOutput", "2004": "clap/modeling_clap.py:ClapTextLayer", "2005": "clap/modeling_clap.py:ClapTextEncoder", "2006": "clap/modeling_clap.py:ClapTextPooler", "2007": "clap/modeling_clap.py:ClapPreTrainedModel", "2008": "clap/modeling_clap.py:ClapAudioModel", "2009": "clap/modeling_clap.py:ClapTextModel", "2010": "clap/modeling_clap.py:ClapModel", "2011": "clap/modeling_clap.py:ClapTextModelWithProjection", "2012": "clap/modeling_clap.py:ClapAudioModelWithProjection", "2013": "electra/modeling_electra.py:ElectraEmbeddings", "2014": "electra/modeling_electra.py:eager_attention_forward", "2015": "electra/modeling_electra.py:ElectraSelfAttention", "2016": "electra/modeling_electra.py:ElectraCrossAttention", "2017": "electra/modeling_electra.py:ElectraSelfOutput", "2018": "electra/modeling_electra.py:ElectraAttention", "2019": "electra/modeling_electra.py:ElectraIntermediate", "2020": "electra/modeling_electra.py:ElectraOutput", "2021": "electra/modeling_electra.py:ElectraLayer", "2022": "electra/modeling_electra.py:ElectraEncoder", "2023": "electra/modeling_electra.py:ElectraDiscriminatorPredictions", "2024": "electra/modeling_electra.py:ElectraGeneratorPredictions", "2025": "electra/modeling_electra.py:ElectraPreTrainedModel", "2026": "electra/modeling_electra.py:ElectraForPreTrainingOutput", "2027": "electra/modeling_electra.py:ElectraModel", "2028": "electra/modeling_electra.py:ElectraClassificationHead", "2029": "electra/modeling_electra.py:ElectraSequenceSummary", "2030": "electra/modeling_electra.py:ElectraForSequenceClassification", "2031": "electra/modeling_electra.py:ElectraForPreTraining", "2032": "electra/modeling_electra.py:ElectraForMaskedLM", "2033": "electra/modeling_electra.py:ElectraForTokenClassification", "2034": "electra/modeling_electra.py:ElectraForQuestionAnswering", "2035": "electra/modeling_electra.py:ElectraForMultipleChoice", "2036": "electra/modeling_electra.py:ElectraForCausalLM", "2037": "glm4v/modeling_glm4v.py:Glm4vRMSNorm", "2038": "glm4v/modeling_glm4v.py:Glm4VisionMlp", "2039": "glm4v/modeling_glm4v.py:Glm4vVisionPatchEmbed", "2040": "glm4v/modeling_glm4v.py:Glm4vVisionRotaryEmbedding", "2041": "glm4v/modeling_glm4v.py:Glm4vVisionPatchMerger", "2042": "glm4v/modeling_glm4v.py:Glm4vVisionEmbeddings", "2043": "glm4v/modeling_glm4v.py:rotate_half", "2044": "glm4v/modeling_glm4v.py:apply_rotary_pos_emb_vision", "2045": "glm4v/modeling_glm4v.py:repeat_kv", "2046": "glm4v/modeling_glm4v.py:eager_attention_forward", "2047": "glm4v/modeling_glm4v.py:Glm4vVisionAttention", "2048": "glm4v/modeling_glm4v.py:Glm4vVisionBlock", "2049": "glm4v/modeling_glm4v.py:Glm4vTextRotaryEmbedding", "2050": "glm4v/modeling_glm4v.py:rotate_half_llm", "2051": "glm4v/modeling_glm4v.py:apply_multimodal_rotary_pos_emb", "2052": "glm4v/modeling_glm4v.py:Glm4vTextAttention", "2053": "glm4v/modeling_glm4v.py:Glm4vTextMLP", "2054": "glm4v/modeling_glm4v.py:Glm4vTextDecoderLayer", "2055": "glm4v/modeling_glm4v.py:Glm4vModelOutputWithPast", "2056": "glm4v/modeling_glm4v.py:Glm4vPreTrainedModel", "2057": "glm4v/modeling_glm4v.py:Glm4vVisionModel", "2058": "glm4v/modeling_glm4v.py:Glm4vTextModel", "2059": "glm4v/modeling_glm4v.py:Glm4vModel", "2060": "glm4v/modeling_glm4v.py:Glm4vCausalLMOutputWithPast", "2061": "glm4v/modeling_glm4v.py:Glm4vForConditionalGeneration", "2062": "exaone4/modeling_exaone4.py:Exaone4RMSNorm", "2063": "exaone4/modeling_exaone4.py:Exaone4RotaryEmbedding", "2064": "exaone4/modeling_exaone4.py:rotate_half", "2065": "exaone4/modeling_exaone4.py:apply_rotary_pos_emb", "2066": "exaone4/modeling_exaone4.py:repeat_kv", "2067": "exaone4/modeling_exaone4.py:eager_attention_forward", "2068": "exaone4/modeling_exaone4.py:Exaone4Attention", "2069": "exaone4/modeling_exaone4.py:Exaone4MLP", "2070": "exaone4/modeling_exaone4.py:Exaone4DecoderLayer", "2071": "exaone4/modeling_exaone4.py:Exaone4PreTrainedModel", "2072": "exaone4/modeling_exaone4.py:Exaone4Model", "2073": "exaone4/modeling_exaone4.py:Exaone4ForCausalLM", "2074": "exaone4/modeling_exaone4.py:Exaone4ForSequenceClassification", "2075": "exaone4/modeling_exaone4.py:Exaone4ForTokenClassification", "2076": "exaone4/modeling_exaone4.py:Exaone4ForQuestionAnswering", "2077": "donut/modeling_donut_swin.py:DonutSwinEncoderOutput", "2078": "donut/modeling_donut_swin.py:DonutSwinModelOutput", "2079": "donut/modeling_donut_swin.py:DonutSwinImageClassifierOutput", "2080": "donut/modeling_donut_swin.py:window_partition", "2081": "donut/modeling_donut_swin.py:window_reverse", "2082": "donut/modeling_donut_swin.py:DonutSwinEmbeddings", "2083": "donut/modeling_donut_swin.py:DonutSwinPatchEmbeddings", "2084": "donut/modeling_donut_swin.py:DonutSwinPatchMerging", "2085": "donut/modeling_donut_swin.py:drop_path", "2086": "donut/modeling_donut_swin.py:DonutSwinDropPath", "2087": "donut/modeling_donut_swin.py:DonutSwinSelfAttention", "2088": "donut/modeling_donut_swin.py:DonutSwinSelfOutput", "2089": "donut/modeling_donut_swin.py:DonutSwinAttention", "2090": "donut/modeling_donut_swin.py:DonutSwinIntermediate", "2091": "donut/modeling_donut_swin.py:DonutSwinOutput", "2092": "donut/modeling_donut_swin.py:DonutSwinLayer", "2093": "donut/modeling_donut_swin.py:DonutSwinStage", "2094": "donut/modeling_donut_swin.py:DonutSwinEncoder", "2095": "donut/modeling_donut_swin.py:DonutSwinPreTrainedModel", "2096": "donut/modeling_donut_swin.py:DonutSwinModel", "2097": "donut/modeling_donut_swin.py:DonutSwinForImageClassification", "2098": "pegasus/modeling_pegasus.py:shift_tokens_right", "2099": "pegasus/modeling_pegasus.py:PegasusSinusoidalPositionalEmbedding", "2100": "pegasus/modeling_pegasus.py:eager_attention_forward", "2101": "pegasus/modeling_pegasus.py:PegasusAttention", "2102": "pegasus/modeling_pegasus.py:PegasusEncoderLayer", "2103": "pegasus/modeling_pegasus.py:PegasusDecoderLayer", "2104": "pegasus/modeling_pegasus.py:PegasusPreTrainedModel", "2105": "pegasus/modeling_pegasus.py:PegasusEncoder", "2106": "pegasus/modeling_pegasus.py:PegasusDecoder", "2107": "pegasus/modeling_pegasus.py:PegasusModel", "2108": "pegasus/modeling_pegasus.py:PegasusForConditionalGeneration", "2109": "pegasus/modeling_pegasus.py:PegasusDecoderWrapper", "2110": "pegasus/modeling_pegasus.py:PegasusForCausalLM", "2111": "longt5/modeling_longt5.py:_pad_to_multiple", "2112": "longt5/modeling_longt5.py:_split_into_blocks", "2113": "longt5/modeling_longt5.py:_concatenate_3_blocks", "2114": "longt5/modeling_longt5.py:_make_3block_relative_position_ids", "2115": "longt5/modeling_longt5.py:_mask_local_attention_mask", "2116": "longt5/modeling_longt5.py:_get_local_attention_mask", "2117": "longt5/modeling_longt5.py:_make_global_fixed_block_ids", "2118": "longt5/modeling_longt5.py:_make_side_relative_position_ids", "2119": "longt5/modeling_longt5.py:_create_global_aggregates", "2120": "longt5/modeling_longt5.py:LongT5LayerNorm", "2121": "longt5/modeling_longt5.py:LongT5DenseActDense", "2122": "longt5/modeling_longt5.py:LongT5DenseGatedActDense", "2123": "longt5/modeling_longt5.py:LongT5LayerFF", "2124": "longt5/modeling_longt5.py:LongT5Attention", "2125": "longt5/modeling_longt5.py:LongT5LocalAttention", "2126": "longt5/modeling_longt5.py:LongT5TransientGlobalAttention", "2127": "longt5/modeling_longt5.py:LongT5LayerSelfAttention", "2128": "longt5/modeling_longt5.py:LongT5LayerLocalSelfAttention", "2129": "longt5/modeling_longt5.py:LongT5LayerTransientGlobalSelfAttention", "2130": "longt5/modeling_longt5.py:LongT5LayerCrossAttention", "2131": "longt5/modeling_longt5.py:LongT5Block", "2132": "longt5/modeling_longt5.py:LongT5PreTrainedModel", "2133": "longt5/modeling_longt5.py:LongT5Stack", "2134": "longt5/modeling_longt5.py:LongT5Model", "2135": "longt5/modeling_longt5.py:LongT5ForConditionalGeneration", "2136": "longt5/modeling_longt5.py:LongT5EncoderModel", "2137": "apertus/modeling_apertus.py:ApertusMLP", "2138": "apertus/modeling_apertus.py:ApertusRMSNorm", "2139": "apertus/modeling_apertus.py:ApertusRotaryEmbedding", "2140": "apertus/modeling_apertus.py:rotate_half", "2141": "apertus/modeling_apertus.py:apply_rotary_pos_emb", "2142": "apertus/modeling_apertus.py:repeat_kv", "2143": "apertus/modeling_apertus.py:eager_attention_forward", "2144": "apertus/modeling_apertus.py:ApertusAttention", "2145": "apertus/modeling_apertus.py:ApertusDecoderLayer", "2146": "apertus/modeling_apertus.py:ApertusPreTrainedModel", "2147": "apertus/modeling_apertus.py:ApertusModel", "2148": "apertus/modeling_apertus.py:ApertusForCausalLM", "2149": "apertus/modeling_apertus.py:ApertusForTokenClassification", "2150": "timesformer/modeling_timesformer.py:TimesformerPatchEmbeddings", "2151": "timesformer/modeling_timesformer.py:TimesformerEmbeddings", "2152": "timesformer/modeling_timesformer.py:drop_path", "2153": "timesformer/modeling_timesformer.py:TimeSformerDropPath", "2154": "timesformer/modeling_timesformer.py:TimesformerSelfAttention", "2155": "timesformer/modeling_timesformer.py:TimesformerSelfOutput", "2156": "timesformer/modeling_timesformer.py:TimeSformerAttention", "2157": "timesformer/modeling_timesformer.py:TimesformerIntermediate", "2158": "timesformer/modeling_timesformer.py:TimesformerOutput", "2159": "timesformer/modeling_timesformer.py:TimesformerLayer", "2160": "timesformer/modeling_timesformer.py:TimesformerEncoder", "2161": "timesformer/modeling_timesformer.py:TimesformerPreTrainedModel", "2162": "timesformer/modeling_timesformer.py:TimesformerModel", "2163": "timesformer/modeling_timesformer.py:TimesformerForVideoClassification", "2164": "nllb_moe/modeling_nllb_moe.py:shift_tokens_right", "2165": "nllb_moe/modeling_nllb_moe.py:load_balancing_loss_func", "2166": "nllb_moe/modeling_nllb_moe.py:NllbMoeScaledWordEmbedding", "2167": "nllb_moe/modeling_nllb_moe.py:NllbMoeSinusoidalPositionalEmbedding", "2168": "nllb_moe/modeling_nllb_moe.py:NllbMoeTop2Router", "2169": "nllb_moe/modeling_nllb_moe.py:NllbMoeDenseActDense", "2170": "nllb_moe/modeling_nllb_moe.py:NllbMoeSparseMLP", "2171": "nllb_moe/modeling_nllb_moe.py:eager_attention_forward", "2172": "nllb_moe/modeling_nllb_moe.py:NllbMoeAttention", "2173": "nllb_moe/modeling_nllb_moe.py:NllbMoeEncoderLayer", "2174": "nllb_moe/modeling_nllb_moe.py:NllbMoeDecoderLayer", "2175": "nllb_moe/modeling_nllb_moe.py:NllbMoePreTrainedModel", "2176": "nllb_moe/modeling_nllb_moe.py:NllbMoeEncoder", "2177": "nllb_moe/modeling_nllb_moe.py:NllbMoeDecoder", "2178": "nllb_moe/modeling_nllb_moe.py:NllbMoeModel", "2179": "nllb_moe/modeling_nllb_moe.py:NllbMoeForConditionalGeneration", "2180": "olmo3/modeling_olmo3.py:Olmo3RMSNorm", "2181": "olmo3/modeling_olmo3.py:repeat_kv", "2182": "olmo3/modeling_olmo3.py:eager_attention_forward", "2183": "olmo3/modeling_olmo3.py:apply_rotary_pos_emb", "2184": "olmo3/modeling_olmo3.py:rotate_half", "2185": "olmo3/modeling_olmo3.py:Olmo3Attention", "2186": "olmo3/modeling_olmo3.py:Olmo3MLP", "2187": "olmo3/modeling_olmo3.py:Olmo3DecoderLayer", "2188": "olmo3/modeling_olmo3.py:Olmo3RotaryEmbedding", "2189": "olmo3/modeling_olmo3.py:Olmo3PreTrainedModel", "2190": "olmo3/modeling_olmo3.py:Olmo3Model", "2191": "olmo3/modeling_olmo3.py:Olmo3ForCausalLM", "2192": "glm4_moe/modeling_glm4_moe.py:repeat_kv", "2193": "glm4_moe/modeling_glm4_moe.py:eager_attention_forward", "2194": "glm4_moe/modeling_glm4_moe.py:rotate_half", "2195": "glm4_moe/modeling_glm4_moe.py:apply_rotary_pos_emb", "2196": "glm4_moe/modeling_glm4_moe.py:Glm4MoeAttention", "2197": "glm4_moe/modeling_glm4_moe.py:Glm4MoeMLP", "2198": "glm4_moe/modeling_glm4_moe.py:Glm4MoeTopkRouter", "2199": "glm4_moe/modeling_glm4_moe.py:Glm4MoeRMSNorm", "2200": "glm4_moe/modeling_glm4_moe.py:Glm4MoeMoE", "2201": "glm4_moe/modeling_glm4_moe.py:Glm4MoeDecoderLayer", "2202": "glm4_moe/modeling_glm4_moe.py:Glm4MoePreTrainedModel", "2203": "glm4_moe/modeling_glm4_moe.py:Glm4MoeRotaryEmbedding", "2204": "glm4_moe/modeling_glm4_moe.py:Glm4MoeModel", "2205": "glm4_moe/modeling_glm4_moe.py:Glm4MoeForCausalLM", "2206": "flex_olmo/modeling_flex_olmo.py:FlexOlmoRMSNorm", "2207": "flex_olmo/modeling_flex_olmo.py:FlexOlmoRotaryEmbedding", "2208": "flex_olmo/modeling_flex_olmo.py:FlexOlmoMLP", "2209": "flex_olmo/modeling_flex_olmo.py:repeat_kv", "2210": "flex_olmo/modeling_flex_olmo.py:eager_attention_forward", "2211": "flex_olmo/modeling_flex_olmo.py:apply_rotary_pos_emb", "2212": "flex_olmo/modeling_flex_olmo.py:rotate_half", "2213": "flex_olmo/modeling_flex_olmo.py:FlexOlmoAttention", "2214": "flex_olmo/modeling_flex_olmo.py:FlexOlmoSparseMoeBlock", "2215": "flex_olmo/modeling_flex_olmo.py:FlexOlmoDecoderLayer", "2216": "flex_olmo/modeling_flex_olmo.py:FlexOlmoPreTrainedModel", "2217": "flex_olmo/modeling_flex_olmo.py:FlexOlmoModel", "2218": "flex_olmo/modeling_flex_olmo.py:load_balancing_loss_func", "2219": "flex_olmo/modeling_flex_olmo.py:FlexOlmoForCausalLM", "2220": "flaubert/modeling_flaubert.py:create_sinusoidal_embeddings", "2221": "flaubert/modeling_flaubert.py:get_masks", "2222": "flaubert/modeling_flaubert.py:MultiHeadAttention", "2223": "flaubert/modeling_flaubert.py:TransformerFFN", "2224": "flaubert/modeling_flaubert.py:FlaubertPredLayer", "2225": "flaubert/modeling_flaubert.py:FlaubertSquadHeadOutput", "2226": "flaubert/modeling_flaubert.py:FlaubertPoolerStartLogits", "2227": "flaubert/modeling_flaubert.py:FlaubertPoolerEndLogits", "2228": "flaubert/modeling_flaubert.py:FlaubertPoolerAnswerClass", "2229": "flaubert/modeling_flaubert.py:FlaubertSQuADHead", "2230": "flaubert/modeling_flaubert.py:FlaubertSequenceSummary", "2231": "flaubert/modeling_flaubert.py:FlaubertPreTrainedModel", "2232": "flaubert/modeling_flaubert.py:FlaubertModel", "2233": "flaubert/modeling_flaubert.py:FlaubertWithLMHeadModel", "2234": "flaubert/modeling_flaubert.py:FlaubertForSequenceClassification", "2235": "flaubert/modeling_flaubert.py:FlaubertForTokenClassification", "2236": "flaubert/modeling_flaubert.py:FlaubertForQuestionAnsweringSimple", "2237": "flaubert/modeling_flaubert.py:FlaubertForQuestionAnsweringOutput", "2238": "flaubert/modeling_flaubert.py:FlaubertForQuestionAnswering", "2239": "flaubert/modeling_flaubert.py:FlaubertForMultipleChoice", "2240": "mobilenet_v2/modeling_mobilenet_v2.py:make_divisible", "2241": "mobilenet_v2/modeling_mobilenet_v2.py:apply_depth_multiplier", "2242": "mobilenet_v2/modeling_mobilenet_v2.py:apply_tf_padding", "2243": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2ConvLayer", "2244": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2InvertedResidual", "2245": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2Stem", "2246": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2PreTrainedModel", "2247": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2Model", "2248": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2ForImageClassification", "2249": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2DeepLabV3Plus", "2250": "mobilenet_v2/modeling_mobilenet_v2.py:MobileNetV2ForSemanticSegmentation", "2251": "openai/modeling_openai.py:Attention", "2252": "openai/modeling_openai.py:MLP", "2253": "openai/modeling_openai.py:Block", "2254": "openai/modeling_openai.py:OpenAIGPTSequenceSummary", "2255": "openai/modeling_openai.py:OpenAIGPTPreTrainedModel", "2256": "openai/modeling_openai.py:OpenAIGPTDoubleHeadsModelOutput", "2257": "openai/modeling_openai.py:OpenAIGPTModel", "2258": "openai/modeling_openai.py:OpenAIGPTLMHeadModel", "2259": "openai/modeling_openai.py:OpenAIGPTDoubleHeadsModel", "2260": "openai/modeling_openai.py:OpenAIGPTForSequenceClassification", "2261": "fuyu/modeling_fuyu.py:FuyuPreTrainedModel", "2262": "fuyu/modeling_fuyu.py:FuyuModel", "2263": "fuyu/modeling_fuyu.py:FuyuForCausalLM", "2264": "bit/modeling_bit.py:get_padding_value", "2265": "bit/modeling_bit.py:WeightStandardizedConv2d", "2266": "bit/modeling_bit.py:BitGroupNormActivation", "2267": "bit/modeling_bit.py:DynamicPad2d", "2268": "bit/modeling_bit.py:BitMaxPool2d", "2269": "bit/modeling_bit.py:BitEmbeddings", "2270": "bit/modeling_bit.py:drop_path", "2271": "bit/modeling_bit.py:BitDropPath", "2272": "bit/modeling_bit.py:make_div", "2273": "bit/modeling_bit.py:BitPreActivationBottleneckLayer", "2274": "bit/modeling_bit.py:BitBottleneckLayer", "2275": "bit/modeling_bit.py:BitDownsampleConv", "2276": "bit/modeling_bit.py:BitStage", "2277": "bit/modeling_bit.py:BitEncoder", "2278": "bit/modeling_bit.py:BitPreTrainedModel", "2279": "bit/modeling_bit.py:BitModel", "2280": "bit/modeling_bit.py:BitForImageClassification", "2281": "bit/modeling_bit.py:BitBackbone", "2282": "vit/modeling_vit.py:ViTEmbeddings", "2283": "vit/modeling_vit.py:ViTPatchEmbeddings", "2284": "vit/modeling_vit.py:eager_attention_forward", "2285": "vit/modeling_vit.py:ViTSelfAttention", "2286": "vit/modeling_vit.py:ViTSelfOutput", "2287": "vit/modeling_vit.py:ViTAttention", "2288": "vit/modeling_vit.py:ViTIntermediate", "2289": "vit/modeling_vit.py:ViTOutput", "2290": "vit/modeling_vit.py:ViTLayer", "2291": "vit/modeling_vit.py:ViTEncoder", "2292": "vit/modeling_vit.py:ViTPreTrainedModel", "2293": "vit/modeling_vit.py:ViTModel", "2294": "vit/modeling_vit.py:ViTPooler", "2295": "vit/modeling_vit.py:ViTForMaskedImageModeling", "2296": "vit/modeling_vit.py:ViTForImageClassification", "2297": "blenderbot/modeling_blenderbot.py:shift_tokens_right", "2298": "blenderbot/modeling_blenderbot.py:BlenderbotLearnedPositionalEmbedding", "2299": "blenderbot/modeling_blenderbot.py:BlenderbotScaledWordEmbedding", "2300": "blenderbot/modeling_blenderbot.py:eager_attention_forward", "2301": "blenderbot/modeling_blenderbot.py:BlenderbotAttention", "2302": "blenderbot/modeling_blenderbot.py:BlenderbotEncoderLayer", "2303": "blenderbot/modeling_blenderbot.py:BlenderbotDecoderLayer", "2304": "blenderbot/modeling_blenderbot.py:BlenderbotPreTrainedModel", "2305": "blenderbot/modeling_blenderbot.py:BlenderbotEncoder", "2306": "blenderbot/modeling_blenderbot.py:BlenderbotDecoder", "2307": "blenderbot/modeling_blenderbot.py:BlenderbotModel", "2308": "blenderbot/modeling_blenderbot.py:BlenderbotForConditionalGeneration", "2309": "blenderbot/modeling_blenderbot.py:BlenderbotDecoderWrapper", "2310": "blenderbot/modeling_blenderbot.py:BlenderbotForCausalLM", "2311": "ernie/modeling_ernie.py:ErnieEmbeddings", "2312": "ernie/modeling_ernie.py:eager_attention_forward", "2313": "ernie/modeling_ernie.py:ErnieSelfAttention", "2314": "ernie/modeling_ernie.py:ErnieCrossAttention", "2315": "ernie/modeling_ernie.py:ErnieSelfOutput", "2316": "ernie/modeling_ernie.py:ErnieAttention", "2317": "ernie/modeling_ernie.py:ErnieIntermediate", "2318": "ernie/modeling_ernie.py:ErnieOutput", "2319": "ernie/modeling_ernie.py:ErnieLayer", "2320": "ernie/modeling_ernie.py:ErniePooler", "2321": "ernie/modeling_ernie.py:ErniePredictionHeadTransform", "2322": "ernie/modeling_ernie.py:ErnieLMPredictionHead", "2323": "ernie/modeling_ernie.py:ErnieEncoder", "2324": "ernie/modeling_ernie.py:ErniePreTrainedModel", "2325": "ernie/modeling_ernie.py:ErnieModel", "2326": "ernie/modeling_ernie.py:ErnieForPreTrainingOutput", "2327": "ernie/modeling_ernie.py:ErniePreTrainingHeads", "2328": "ernie/modeling_ernie.py:ErnieForPreTraining", "2329": "ernie/modeling_ernie.py:ErnieOnlyMLMHead", "2330": "ernie/modeling_ernie.py:ErnieForCausalLM", "2331": "ernie/modeling_ernie.py:ErnieForMaskedLM", "2332": "ernie/modeling_ernie.py:ErnieOnlyNSPHead", "2333": "ernie/modeling_ernie.py:ErnieForNextSentencePrediction", "2334": "ernie/modeling_ernie.py:ErnieForSequenceClassification", "2335": "ernie/modeling_ernie.py:ErnieForMultipleChoice", "2336": "ernie/modeling_ernie.py:ErnieForTokenClassification", "2337": "ernie/modeling_ernie.py:ErnieForQuestionAnswering", "2338": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrDecoderOutput", "2339": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrModelOutput", "2340": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrObjectDetectionOutput", "2341": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrSegmentationOutput", "2342": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrFrozenBatchNorm2d", "2343": "conditional_detr/modeling_conditional_detr.py:replace_batch_norm", "2344": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrConvEncoder", "2345": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrConvModel", "2346": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrSinePositionEmbedding", "2347": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrLearnedPositionEmbedding", "2348": "conditional_detr/modeling_conditional_detr.py:build_position_encoding", "2349": "conditional_detr/modeling_conditional_detr.py:gen_sine_position_embeddings", "2350": "conditional_detr/modeling_conditional_detr.py:inverse_sigmoid", "2351": "conditional_detr/modeling_conditional_detr.py:DetrAttention", "2352": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrAttention", "2353": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrEncoderLayer", "2354": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrDecoderLayer", "2355": "conditional_detr/modeling_conditional_detr.py:MLP", "2356": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrPreTrainedModel", "2357": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrEncoder", "2358": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrDecoder", "2359": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrModel", "2360": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrMLPPredictionHead", "2361": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrForObjectDetection", "2362": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrForSegmentation", "2363": "conditional_detr/modeling_conditional_detr.py:_expand", "2364": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrMaskHeadSmallConv", "2365": "conditional_detr/modeling_conditional_detr.py:ConditionalDetrMHAttentionMap", "2366": "focalnet/modeling_focalnet.py:FocalNetEncoderOutput", "2367": "focalnet/modeling_focalnet.py:FocalNetModelOutput", "2368": "focalnet/modeling_focalnet.py:FocalNetMaskedImageModelingOutput", "2369": "focalnet/modeling_focalnet.py:FocalNetImageClassifierOutput", "2370": "focalnet/modeling_focalnet.py:FocalNetEmbeddings", "2371": "focalnet/modeling_focalnet.py:FocalNetPatchEmbeddings", "2372": "focalnet/modeling_focalnet.py:drop_path", "2373": "focalnet/modeling_focalnet.py:FocalNetDropPath", "2374": "focalnet/modeling_focalnet.py:FocalNetModulation", "2375": "focalnet/modeling_focalnet.py:FocalNetMlp", "2376": "focalnet/modeling_focalnet.py:FocalNetLayer", "2377": "focalnet/modeling_focalnet.py:FocalNetStage", "2378": "focalnet/modeling_focalnet.py:FocalNetEncoder", "2379": "focalnet/modeling_focalnet.py:FocalNetPreTrainedModel", "2380": "focalnet/modeling_focalnet.py:FocalNetModel", "2381": "focalnet/modeling_focalnet.py:FocalNetForMaskedImageModeling", "2382": "focalnet/modeling_focalnet.py:FocalNetForImageClassification", "2383": "focalnet/modeling_focalnet.py:FocalNetBackbone", "2384": "mamba2/modeling_mamba2.py:pad_tensor_by_size", "2385": "mamba2/modeling_mamba2.py:reshape_into_chunks", "2386": "mamba2/modeling_mamba2.py:segment_sum", "2387": "mamba2/modeling_mamba2.py:apply_mask_to_padding_states", "2388": "mamba2/modeling_mamba2.py:Mamba2Cache", "2389": "mamba2/modeling_mamba2.py:MambaRMSNormGated", "2390": "mamba2/modeling_mamba2.py:Mamba2Mixer", "2391": "mamba2/modeling_mamba2.py:Mamba2RMSNorm", "2392": "mamba2/modeling_mamba2.py:Mamba2Block", "2393": "mamba2/modeling_mamba2.py:Mamba2PreTrainedModel", "2394": "mamba2/modeling_mamba2.py:Mamba2Output", "2395": "mamba2/modeling_mamba2.py:Mamba2CausalLMOutput", "2396": "mamba2/modeling_mamba2.py:Mamba2Model", "2397": "mamba2/modeling_mamba2.py:Mamba2ForCausalLM", "2398": "mvp/modeling_mvp.py:shift_tokens_right", "2399": "mvp/modeling_mvp.py:MvpLearnedPositionalEmbedding", "2400": "mvp/modeling_mvp.py:MvpAttention", "2401": "mvp/modeling_mvp.py:MvpEncoderLayer", "2402": "mvp/modeling_mvp.py:MvpDecoderLayer", "2403": "mvp/modeling_mvp.py:MvpClassificationHead", "2404": "mvp/modeling_mvp.py:MvpPrompt", "2405": "mvp/modeling_mvp.py:MvpPreTrainedModel", "2406": "mvp/modeling_mvp.py:MvpEncoder", "2407": "mvp/modeling_mvp.py:MvpDecoder", "2408": "mvp/modeling_mvp.py:MvpModel", "2409": "mvp/modeling_mvp.py:MvpForConditionalGeneration", "2410": "mvp/modeling_mvp.py:MvpForSequenceClassification", "2411": "mvp/modeling_mvp.py:MvpForQuestionAnswering", "2412": "mvp/modeling_mvp.py:MvpDecoderWrapper", "2413": "mvp/modeling_mvp.py:MvpForCausalLM", "2414": "kosmos2/modeling_kosmos2.py:_expand_mask", "2415": "kosmos2/modeling_kosmos2.py:_make_causal_mask", "2416": "kosmos2/modeling_kosmos2.py:Kosmos2ModelOutput", "2417": "kosmos2/modeling_kosmos2.py:Kosmos2ForConditionalGenerationModelOutput", "2418": "kosmos2/modeling_kosmos2.py:Kosmos2VisionEmbeddings", "2419": "kosmos2/modeling_kosmos2.py:eager_attention_forward", "2420": "kosmos2/modeling_kosmos2.py:Kosmos2VisionAttention", "2421": "kosmos2/modeling_kosmos2.py:Kosmos2VisionMLP", "2422": "kosmos2/modeling_kosmos2.py:Kosmos2VisionEncoderLayer", "2423": "kosmos2/modeling_kosmos2.py:Kosmos2VisionEncoder", "2424": "kosmos2/modeling_kosmos2.py:Kosmos2VisionTransformer", "2425": "kosmos2/modeling_kosmos2.py:Kosmos2TextSinusoidalPositionalEmbedding", "2426": "kosmos2/modeling_kosmos2.py:KosmosTextAttention", "2427": "kosmos2/modeling_kosmos2.py:Kosmos2TextFFN", "2428": "kosmos2/modeling_kosmos2.py:Kosmos2TextBlock", "2429": "kosmos2/modeling_kosmos2.py:Kosmos2TextTransformer", "2430": "kosmos2/modeling_kosmos2.py:Kosmos2PreTrainedModel", "2431": "kosmos2/modeling_kosmos2.py:Kosmos2VisionModel", "2432": "kosmos2/modeling_kosmos2.py:Kosmos2TextModel", "2433": "kosmos2/modeling_kosmos2.py:Kosmos2TextForCausalLM", "2434": "kosmos2/modeling_kosmos2.py:Kosmos2ImageToTextProjection", "2435": "kosmos2/modeling_kosmos2.py:Kosmos2Model", "2436": "kosmos2/modeling_kosmos2.py:Kosmos2ForConditionalGeneration", "2437": "grounding_dino/modeling_grounding_dino.py:MultiScaleDeformableAttention", "2438": "grounding_dino/modeling_grounding_dino.py:GroundingDinoDecoderOutput", "2439": "grounding_dino/modeling_grounding_dino.py:GroundingDinoEncoderOutput", "2440": "grounding_dino/modeling_grounding_dino.py:GroundingDinoModelOutput", "2441": "grounding_dino/modeling_grounding_dino.py:GroundingDinoObjectDetectionOutput", "2442": "grounding_dino/modeling_grounding_dino.py:GroundingDinoFrozenBatchNorm2d", "2443": "grounding_dino/modeling_grounding_dino.py:replace_batch_norm", "2444": "grounding_dino/modeling_grounding_dino.py:GroundingDinoConvEncoder", "2445": "grounding_dino/modeling_grounding_dino.py:GroundingDinoConvModel", "2446": "grounding_dino/modeling_grounding_dino.py:GroundingDinoSinePositionEmbedding", "2447": "grounding_dino/modeling_grounding_dino.py:GroundingDinoLearnedPositionEmbedding", "2448": "grounding_dino/modeling_grounding_dino.py:build_position_encoding", "2449": "grounding_dino/modeling_grounding_dino.py:GroundingDinoMultiscaleDeformableAttention", "2450": "grounding_dino/modeling_grounding_dino.py:GroundingDinoTextEnhancerLayer", "2451": "grounding_dino/modeling_grounding_dino.py:GroundingDinoBiMultiHeadAttention", "2452": "grounding_dino/modeling_grounding_dino.py:drop_path", "2453": "grounding_dino/modeling_grounding_dino.py:GroundingDinoDropPath", "2454": "grounding_dino/modeling_grounding_dino.py:GroundingDinoFusionLayer", "2455": "grounding_dino/modeling_grounding_dino.py:GroundingDinoDeformableLayer", "2456": "grounding_dino/modeling_grounding_dino.py:get_sine_pos_embed", "2457": "grounding_dino/modeling_grounding_dino.py:GroundingDinoEncoderLayer", "2458": "grounding_dino/modeling_grounding_dino.py:GroundingDinoMultiheadAttention", "2459": "grounding_dino/modeling_grounding_dino.py:GroundingDinoDecoderLayer", "2460": "grounding_dino/modeling_grounding_dino.py:GroundingDinoContrastiveEmbedding", "2461": "grounding_dino/modeling_grounding_dino.py:GroundingDinoPreTrainedModel", "2462": "grounding_dino/modeling_grounding_dino.py:GroundingDinoEncoder", "2463": "grounding_dino/modeling_grounding_dino.py:GroundingDinoDecoder", "2464": "grounding_dino/modeling_grounding_dino.py:generate_masks_with_special_tokens_and_transfer_map", "2465": "grounding_dino/modeling_grounding_dino.py:GroundingDinoModel", "2466": "grounding_dino/modeling_grounding_dino.py:GroundingDinoMLPPredictionHead", "2467": "grounding_dino/modeling_grounding_dino.py:build_label_maps", "2468": "grounding_dino/modeling_grounding_dino.py:build_text_mask", "2469": "grounding_dino/modeling_grounding_dino.py:GroundingDinoForObjectDetection", "2470": "bros/modeling_bros.py:BrosSpadeOutput", "2471": "bros/modeling_bros.py:BrosPositionalEmbedding1D", "2472": "bros/modeling_bros.py:BrosPositionalEmbedding2D", "2473": "bros/modeling_bros.py:BrosBboxEmbeddings", "2474": "bros/modeling_bros.py:BrosTextEmbeddings", "2475": "bros/modeling_bros.py:BrosSelfAttention", "2476": "bros/modeling_bros.py:BrosSelfOutput", "2477": "bros/modeling_bros.py:BrosAttention", "2478": "bros/modeling_bros.py:BrosIntermediate", "2479": "bros/modeling_bros.py:BrosOutput", "2480": "bros/modeling_bros.py:BrosLayer", "2481": "bros/modeling_bros.py:BrosEncoder", "2482": "bros/modeling_bros.py:BrosPooler", "2483": "bros/modeling_bros.py:BrosRelationExtractor", "2484": "bros/modeling_bros.py:BrosPreTrainedModel", "2485": "bros/modeling_bros.py:BrosModel", "2486": "bros/modeling_bros.py:BrosForTokenClassification", "2487": "bros/modeling_bros.py:BrosSpadeEEForTokenClassification", "2488": "bros/modeling_bros.py:BrosSpadeELForTokenClassification", "2489": "qwen3/modeling_qwen3.py:Qwen3RMSNorm", "2490": "qwen3/modeling_qwen3.py:Qwen3MLP", "2491": "qwen3/modeling_qwen3.py:rotate_half", "2492": "qwen3/modeling_qwen3.py:apply_rotary_pos_emb", "2493": "qwen3/modeling_qwen3.py:repeat_kv", "2494": "qwen3/modeling_qwen3.py:eager_attention_forward", "2495": "qwen3/modeling_qwen3.py:Qwen3Attention", "2496": "qwen3/modeling_qwen3.py:Qwen3DecoderLayer", "2497": "qwen3/modeling_qwen3.py:Qwen3PreTrainedModel", "2498": "qwen3/modeling_qwen3.py:Qwen3RotaryEmbedding", "2499": "qwen3/modeling_qwen3.py:Qwen3Model", "2500": "qwen3/modeling_qwen3.py:Qwen3ForCausalLM", "2501": "qwen3/modeling_qwen3.py:Qwen3ForSequenceClassification", "2502": "qwen3/modeling_qwen3.py:Qwen3ForTokenClassification", "2503": "qwen3/modeling_qwen3.py:Qwen3ForQuestionAnswering", "2504": "idefics/modeling_idefics.py:IdeficsBaseModelOutputWithPast", "2505": "idefics/modeling_idefics.py:IdeficsCausalLMOutputWithPast", "2506": "idefics/modeling_idefics.py:expand_inputs_for_generation", "2507": "idefics/modeling_idefics.py:freeze_model", "2508": "idefics/modeling_idefics.py:IdeficsDecoupledEmbedding", "2509": "idefics/modeling_idefics.py:IdeficsDecoupledLinear", "2510": "idefics/modeling_idefics.py:IdeficsRMSNorm", "2511": "idefics/modeling_idefics.py:IdeficsEmbedding", "2512": "idefics/modeling_idefics.py:rotate_half", "2513": "idefics/modeling_idefics.py:apply_rotary_pos_emb", "2514": "idefics/modeling_idefics.py:IdeficsMLP", "2515": "idefics/modeling_idefics.py:eager_attention_forward", "2516": "idefics/modeling_idefics.py:IdeficsAttention", "2517": "idefics/modeling_idefics.py:IdeficsDecoderLayer", "2518": "idefics/modeling_idefics.py:IdeficsGatedCrossAttentionLayer", "2519": "idefics/modeling_idefics.py:IdeficsPreTrainedModel", "2520": "idefics/modeling_idefics.py:IdeficsModel", "2521": "idefics/modeling_idefics.py:IdeficsForVisionText2Text", "2522": "phimoe/modeling_phimoe.py:load_balancing_loss_func", "2523": "phimoe/modeling_phimoe.py:PhimoeRotaryEmbedding", "2524": "phimoe/modeling_phimoe.py:rotate_half", "2525": "phimoe/modeling_phimoe.py:apply_rotary_pos_emb", "2526": "phimoe/modeling_phimoe.py:repeat_kv", "2527": "phimoe/modeling_phimoe.py:PhimoeAttention", "2528": "phimoe/modeling_phimoe.py:PhimoeFlashAttention2", "2529": "phimoe/modeling_phimoe.py:PhimoeSdpaAttention", "2530": "phimoe/modeling_phimoe.py:PhimoeBlockSparseTop2MLP", "2531": "phimoe/modeling_phimoe.py:MultiplierProcessor", "2532": "phimoe/modeling_phimoe.py:sparsemixer", "2533": "phimoe/modeling_phimoe.py:PhimoeSparseMoeBlock", "2534": "phimoe/modeling_phimoe.py:PhimoeDecoderLayer", "2535": "phimoe/modeling_phimoe.py:PhimoePreTrainedModel", "2536": "phimoe/modeling_phimoe.py:PhimoeModel", "2537": "phimoe/modeling_phimoe.py:PhimoeForCausalLM", "2538": "phimoe/modeling_phimoe.py:PhimoeForSequenceClassification", "2539": "pvt_v2/modeling_pvt_v2.py:drop_path", "2540": "pvt_v2/modeling_pvt_v2.py:PvtV2DropPath", "2541": "pvt_v2/modeling_pvt_v2.py:PvtV2OverlapPatchEmbeddings", "2542": "pvt_v2/modeling_pvt_v2.py:PvtV2DepthWiseConv", "2543": "pvt_v2/modeling_pvt_v2.py:PvtV2SelfAttention", "2544": "pvt_v2/modeling_pvt_v2.py:PvtV2ConvFeedForwardNetwork", "2545": "pvt_v2/modeling_pvt_v2.py:PvtV2BlockLayer", "2546": "pvt_v2/modeling_pvt_v2.py:PvtV2EncoderLayer", "2547": "pvt_v2/modeling_pvt_v2.py:PvtV2Encoder", "2548": "pvt_v2/modeling_pvt_v2.py:PvtV2PreTrainedModel", "2549": "pvt_v2/modeling_pvt_v2.py:PvtV2Model", "2550": "pvt_v2/modeling_pvt_v2.py:PvtV2ForImageClassification", "2551": "pvt_v2/modeling_pvt_v2.py:PvtV2Backbone", "2552": "llava_onevision/modeling_llava_onevision.py:LlavaOnevisionModelOutputWithPast", "2553": "llava_onevision/modeling_llava_onevision.py:LlavaOnevisionCausalLMOutputWithPast", "2554": "llava_onevision/modeling_llava_onevision.py:LlavaOnevisionPreTrainedModel", "2555": "llava_onevision/modeling_llava_onevision.py:LlavaOnevisionMultiModalProjector", "2556": "llava_onevision/modeling_llava_onevision.py:get_anyres_image_grid_shape", "2557": "llava_onevision/modeling_llava_onevision.py:image_size_to_num_patches", "2558": "llava_onevision/modeling_llava_onevision.py:unpad_image", "2559": "llava_onevision/modeling_llava_onevision.py:LlavaOnevisionModel", "2560": "llava_onevision/modeling_llava_onevision.py:LlavaOnevisionForConditionalGeneration", "2561": "vipllava/modeling_vipllava.py:VipLlavaModelOutputWithPast", "2562": "vipllava/modeling_vipllava.py:VipLlavaCausalLMOutputWithPast", "2563": "vipllava/modeling_vipllava.py:VipLlavaMultiModalProjector", "2564": "vipllava/modeling_vipllava.py:VipLlavaPreTrainedModel", "2565": "vipllava/modeling_vipllava.py:VipLlavaModel", "2566": "vipllava/modeling_vipllava.py:VipLlavaForConditionalGeneration", "2567": "pix2struct/modeling_pix2struct.py:Pix2StructLayerNorm", "2568": "pix2struct/modeling_pix2struct.py:Pix2StructVisionEmbeddings", "2569": "pix2struct/modeling_pix2struct.py:Pix2StructVisionAttention", "2570": "pix2struct/modeling_pix2struct.py:Pix2StructVisionMlp", "2571": "pix2struct/modeling_pix2struct.py:Pix2StructVisionLayer", "2572": "pix2struct/modeling_pix2struct.py:Pix2StructVisionEncoder", "2573": "pix2struct/modeling_pix2struct.py:Pix2StructPreTrainedModel", "2574": "pix2struct/modeling_pix2struct.py:Pix2StructVisionModel", "2575": "pix2struct/modeling_pix2struct.py:Pix2StructTextDenseGatedActDense", "2576": "pix2struct/modeling_pix2struct.py:Pix2StructTextLayerFF", "2577": "pix2struct/modeling_pix2struct.py:Pix2StructTextAttention", "2578": "pix2struct/modeling_pix2struct.py:Pix2StructTextLayerSelfAttention", "2579": "pix2struct/modeling_pix2struct.py:Pix2StructTextLayerCrossAttention", "2580": "pix2struct/modeling_pix2struct.py:Pix2StructTextBlock", "2581": "pix2struct/modeling_pix2struct.py:Pix2StructTextModel", "2582": "pix2struct/modeling_pix2struct.py:Pix2StructForConditionalGeneration", "2583": "mobilevitv2/modeling_mobilevitv2.py:make_divisible", "2584": "mobilevitv2/modeling_mobilevitv2.py:clip", "2585": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2ConvLayer", "2586": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2InvertedResidual", "2587": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2MobileNetLayer", "2588": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2LinearSelfAttention", "2589": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2FFN", "2590": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2TransformerLayer", "2591": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2Transformer", "2592": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2Layer", "2593": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2Encoder", "2594": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2PreTrainedModel", "2595": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2Model", "2596": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2ForImageClassification", "2597": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2ASPPPooling", "2598": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2ASPP", "2599": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2DeepLabV3", "2600": "mobilevitv2/modeling_mobilevitv2.py:MobileViTV2ForSemanticSegmentation", "2601": "deformable_detr/modeling_deformable_detr.py:MultiScaleDeformableAttention", "2602": "deformable_detr/modeling_deformable_detr.py:DeformableDetrDecoderOutput", "2603": "deformable_detr/modeling_deformable_detr.py:DeformableDetrModelOutput", "2604": "deformable_detr/modeling_deformable_detr.py:DeformableDetrObjectDetectionOutput", "2605": "deformable_detr/modeling_deformable_detr.py:_get_clones", "2606": "deformable_detr/modeling_deformable_detr.py:inverse_sigmoid", "2607": "deformable_detr/modeling_deformable_detr.py:DeformableDetrFrozenBatchNorm2d", "2608": "deformable_detr/modeling_deformable_detr.py:replace_batch_norm", "2609": "deformable_detr/modeling_deformable_detr.py:DeformableDetrConvEncoder", "2610": "deformable_detr/modeling_deformable_detr.py:DeformableDetrConvModel", "2611": "deformable_detr/modeling_deformable_detr.py:DeformableDetrSinePositionEmbedding", "2612": "deformable_detr/modeling_deformable_detr.py:DeformableDetrLearnedPositionEmbedding", "2613": "deformable_detr/modeling_deformable_detr.py:build_position_encoding", "2614": "deformable_detr/modeling_deformable_detr.py:DeformableDetrMultiscaleDeformableAttention", "2615": "deformable_detr/modeling_deformable_detr.py:DeformableDetrMultiheadAttention", "2616": "deformable_detr/modeling_deformable_detr.py:DeformableDetrEncoderLayer", "2617": "deformable_detr/modeling_deformable_detr.py:DeformableDetrDecoderLayer", "2618": "deformable_detr/modeling_deformable_detr.py:DeformableDetrPreTrainedModel", "2619": "deformable_detr/modeling_deformable_detr.py:DeformableDetrEncoder", "2620": "deformable_detr/modeling_deformable_detr.py:DeformableDetrDecoder", "2621": "deformable_detr/modeling_deformable_detr.py:DeformableDetrModel", "2622": "deformable_detr/modeling_deformable_detr.py:DeformableDetrMLPPredictionHead", "2623": "deformable_detr/modeling_deformable_detr.py:DeformableDetrForObjectDetection", "2624": "encoder_decoder/modeling_encoder_decoder.py:shift_tokens_right", "2625": "encoder_decoder/modeling_encoder_decoder.py:EncoderDecoderModel", "2626": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:GPTNeoXJapanesePreTrainedModel", "2627": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:GPTNeoXJapaneseAttention", "2628": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:GPTNeoXJapaneseRotaryEmbedding", "2629": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:rotate_half", "2630": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:apply_rotary_pos_emb", "2631": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:bias_dropout_add", "2632": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:GPTNeoXJapaneseMLP", "2633": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:GPTNeoXJapaneseLayer", "2634": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:GPTNeoXJapaneseModel", "2635": "gpt_neox_japanese/modeling_gpt_neox_japanese.py:GPTNeoXJapaneseForCausalLM", "2636": "videomae/modeling_videomae.py:VideoMAEDecoderOutput", "2637": "videomae/modeling_videomae.py:VideoMAEForPreTrainingOutput", "2638": "videomae/modeling_videomae.py:get_sinusoid_encoding_table", "2639": "videomae/modeling_videomae.py:VideoMAEEmbeddings", "2640": "videomae/modeling_videomae.py:VideoMAEPatchEmbeddings", "2641": "videomae/modeling_videomae.py:eager_attention_forward", "2642": "videomae/modeling_videomae.py:VideoMAESelfAttention", "2643": "videomae/modeling_videomae.py:VideoMAESelfOutput", "2644": "videomae/modeling_videomae.py:VideoMAEAttention", "2645": "videomae/modeling_videomae.py:VideoMAEIntermediate", "2646": "videomae/modeling_videomae.py:VideoMAEOutput", "2647": "videomae/modeling_videomae.py:VideoMAELayer", "2648": "videomae/modeling_videomae.py:VideoMAEEncoder", "2649": "videomae/modeling_videomae.py:VideoMAEPreTrainedModel", "2650": "videomae/modeling_videomae.py:VideoMAEModel", "2651": "videomae/modeling_videomae.py:VideoMAEDecoder", "2652": "videomae/modeling_videomae.py:VideoMAEForPreTraining", "2653": "videomae/modeling_videomae.py:VideoMAEForVideoClassification", "2654": "regnet/modeling_regnet.py:RegNetConvLayer", "2655": "regnet/modeling_regnet.py:RegNetEmbeddings", "2656": "regnet/modeling_regnet.py:RegNetShortCut", "2657": "regnet/modeling_regnet.py:RegNetSELayer", "2658": "regnet/modeling_regnet.py:RegNetXLayer", "2659": "regnet/modeling_regnet.py:RegNetYLayer", "2660": "regnet/modeling_regnet.py:RegNetStage", "2661": "regnet/modeling_regnet.py:RegNetEncoder", "2662": "regnet/modeling_regnet.py:RegNetPreTrainedModel", "2663": "regnet/modeling_regnet.py:RegNetModel", "2664": "regnet/modeling_regnet.py:RegNetForImageClassification", "2665": "luke/modeling_luke.py:BaseLukeModelOutputWithPooling", "2666": "luke/modeling_luke.py:BaseLukeModelOutput", "2667": "luke/modeling_luke.py:LukeMaskedLMOutput", "2668": "luke/modeling_luke.py:EntityClassificationOutput", "2669": "luke/modeling_luke.py:EntityPairClassificationOutput", "2670": "luke/modeling_luke.py:EntitySpanClassificationOutput", "2671": "luke/modeling_luke.py:LukeSequenceClassifierOutput", "2672": "luke/modeling_luke.py:LukeTokenClassifierOutput", "2673": "luke/modeling_luke.py:LukeQuestionAnsweringModelOutput", "2674": "luke/modeling_luke.py:LukeMultipleChoiceModelOutput", "2675": "luke/modeling_luke.py:LukeEmbeddings", "2676": "luke/modeling_luke.py:LukeEntityEmbeddings", "2677": "luke/modeling_luke.py:LukeSelfAttention", "2678": "luke/modeling_luke.py:LukeSelfOutput", "2679": "luke/modeling_luke.py:LukeAttention", "2680": "luke/modeling_luke.py:LukeIntermediate", "2681": "luke/modeling_luke.py:LukeOutput", "2682": "luke/modeling_luke.py:LukeLayer", "2683": "luke/modeling_luke.py:LukeEncoder", "2684": "luke/modeling_luke.py:LukePooler", "2685": "luke/modeling_luke.py:EntityPredictionHeadTransform", "2686": "luke/modeling_luke.py:EntityPredictionHead", "2687": "luke/modeling_luke.py:LukePreTrainedModel", "2688": "luke/modeling_luke.py:LukeModel", "2689": "luke/modeling_luke.py:create_position_ids_from_input_ids", "2690": "luke/modeling_luke.py:LukeLMHead", "2691": "luke/modeling_luke.py:LukeForMaskedLM", "2692": "luke/modeling_luke.py:LukeForEntityClassification", "2693": "luke/modeling_luke.py:LukeForEntityPairClassification", "2694": "luke/modeling_luke.py:LukeForEntitySpanClassification", "2695": "luke/modeling_luke.py:LukeForSequenceClassification", "2696": "luke/modeling_luke.py:LukeForTokenClassification", "2697": "luke/modeling_luke.py:LukeForQuestionAnswering", "2698": "luke/modeling_luke.py:LukeForMultipleChoice", "2699": "perception_lm/modeling_perception_lm.py:PerceptionLMAdaptiveAvgPooling", "2700": "perception_lm/modeling_perception_lm.py:PerceptionLMMultiModalProjector", "2701": "perception_lm/modeling_perception_lm.py:PerceptionLMPreTrainedModel", "2702": "perception_lm/modeling_perception_lm.py:PerceptionLMModelOutputWithPast", "2703": "perception_lm/modeling_perception_lm.py:PerceptionLMCausalLMOutputWithPast", "2704": "perception_lm/modeling_perception_lm.py:PerceptionLMModel", "2705": "perception_lm/modeling_perception_lm.py:PerceptionLMForConditionalGeneration", "2706": "segformer/modeling_segformer.py:SegFormerImageClassifierOutput", "2707": "segformer/modeling_segformer.py:drop_path", "2708": "segformer/modeling_segformer.py:SegformerDropPath", "2709": "segformer/modeling_segformer.py:SegformerOverlapPatchEmbeddings", "2710": "segformer/modeling_segformer.py:SegformerEfficientSelfAttention", "2711": "segformer/modeling_segformer.py:SegformerSelfOutput", "2712": "segformer/modeling_segformer.py:SegformerAttention", "2713": "segformer/modeling_segformer.py:SegformerDWConv", "2714": "segformer/modeling_segformer.py:SegformerMixFFN", "2715": "segformer/modeling_segformer.py:SegformerLayer", "2716": "segformer/modeling_segformer.py:SegformerEncoder", "2717": "segformer/modeling_segformer.py:SegformerPreTrainedModel", "2718": "segformer/modeling_segformer.py:SegformerModel", "2719": "segformer/modeling_segformer.py:SegformerForImageClassification", "2720": "segformer/modeling_segformer.py:SegformerMLP", "2721": "segformer/modeling_segformer.py:SegformerDecodeHead", "2722": "segformer/modeling_segformer.py:SegformerForSemanticSegmentation", "2723": "wavlm/modeling_wavlm.py:WavLMSamePadLayer", "2724": "wavlm/modeling_wavlm.py:WavLMPositionalConvEmbedding", "2725": "wavlm/modeling_wavlm.py:WavLMFeatureProjection", "2726": "wavlm/modeling_wavlm.py:WavLMAttention", "2727": "wavlm/modeling_wavlm.py:WavLMFeedForward", "2728": "wavlm/modeling_wavlm.py:WavLMEncoderLayer", "2729": "wavlm/modeling_wavlm.py:WavLMEncoderLayerStableLayerNorm", "2730": "wavlm/modeling_wavlm.py:WavLMEncoder", "2731": "wavlm/modeling_wavlm.py:WavLMEncoderStableLayerNorm", "2732": "wavlm/modeling_wavlm.py:WavLMGumbelVectorQuantizer", "2733": "wavlm/modeling_wavlm.py:WavLMPreTrainedModel", "2734": "wavlm/modeling_wavlm.py:WavLMNoLayerNormConvLayer", "2735": "wavlm/modeling_wavlm.py:WavLMLayerNormConvLayer", "2736": "wavlm/modeling_wavlm.py:WavLMGroupNormConvLayer", "2737": "wavlm/modeling_wavlm.py:WavLMFeatureEncoder", "2738": "wavlm/modeling_wavlm.py:WavLMAdapterLayer", "2739": "wavlm/modeling_wavlm.py:WavLMAdapter", "2740": "wavlm/modeling_wavlm.py:_compute_mask_indices", "2741": "wavlm/modeling_wavlm.py:WavLMModel", "2742": "wavlm/modeling_wavlm.py:WavLMForCTC", "2743": "wavlm/modeling_wavlm.py:WavLMForSequenceClassification", "2744": "wavlm/modeling_wavlm.py:WavLMForAudioFrameClassification", "2745": "wavlm/modeling_wavlm.py:AMSoftmaxLoss", "2746": "wavlm/modeling_wavlm.py:TDNNLayer", "2747": "wavlm/modeling_wavlm.py:WavLMForXVector", "2748": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoePreTrainedModel", "2749": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:_get_feat_extract_output_lengths", "2750": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoePreTrainedModelForConditionalGeneration", "2751": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:repeat_kv", "2752": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:eager_attention_forward", "2753": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeAudioAttention", "2754": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeAudioEncoderLayer", "2755": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:SinusoidsPositionEmbedding", "2756": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeAudioEncoder", "2757": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:rotate_half", "2758": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:apply_rotary_pos_emb_vision", "2759": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeVisionAttention", "2760": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeVisionPatchMerger", "2761": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeVisionMLP", "2762": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeVisionPatchEmbed", "2763": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeVisionRotaryEmbedding", "2764": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeVisionBlock", "2765": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeVisionEncoder", "2766": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextRotaryEmbedding", "2767": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextMLP", "2768": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextSparseMoeBlock", "2769": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextRMSNorm", "2770": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:apply_rotary_pos_emb", "2771": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextAttention", "2772": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextDecoderLayer", "2773": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextPreTrainedModel", "2774": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTextRMSNorm", "2775": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerTextModel", "2776": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerCausalLMOutputWithPast", "2777": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:load_balancing_loss_func", "2778": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeThinkerForConditionalGeneration", "2779": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerResizeMLP", "2780": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerCodePredictorOutputWithPast", "2781": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeRMSNorm", "2782": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerCodePredictorAttention", "2783": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeMLP", "2784": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerCodePredictorDecoderLayer", "2785": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeRotaryEmbedding", "2786": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerCodePredictorModel", "2787": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerCodePredictorModelForConditionalGeneration", "2788": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerOutputWithPast", "2789": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerRotaryEmbedding", "2790": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerTextMLP", "2791": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerTextSparseMoeBlock", "2792": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerDecoderLayer", "2793": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerModel", "2794": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeTalkerForConditionalGeneration", "2795": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCausalConvNet", "2796": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCausalTransConvNet", "2797": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeConvNeXtBlock", "2798": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavRotatoryEmbedding", "2799": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavAttention", "2800": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavMlp", "2801": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavRMSNorm", "2802": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavLayerScale", "2803": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavTransformerLayer", "2804": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavTransformerModel", "2805": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:SnakeBeta", "2806": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavDecoderResidualUnit", "2807": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2WavDecoderBlock", "2808": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeCode2Wav", "2809": "qwen3_omni_moe/modeling_qwen3_omni_moe.py:Qwen3OmniMoeForConditionalGeneration", "2810": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormEmbeddings", "2811": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:eager_attention_forward", "2812": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormSelfAttention", "2813": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormCrossAttention", "2814": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormSelfOutput", "2815": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormAttention", "2816": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormIntermediate", "2817": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormOutput", "2818": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormLayer", "2819": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormEncoder", "2820": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormPooler", "2821": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormPreTrainedModel", "2822": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormModel", "2823": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormForCausalLM", "2824": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormForMaskedLM", "2825": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormLMHead", "2826": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormForSequenceClassification", "2827": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormForMultipleChoice", "2828": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormForTokenClassification", "2829": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormClassificationHead", "2830": "roberta_prelayernorm/modeling_roberta_prelayernorm.py:RobertaPreLayerNormForQuestionAnswering", "2831": "univnet/modeling_univnet.py:UnivNetModelOutput", "2832": "univnet/modeling_univnet.py:UnivNetKernelPredictorResidualBlock", "2833": "univnet/modeling_univnet.py:UnivNetKernelPredictor", "2834": "univnet/modeling_univnet.py:UnivNetLvcResidualBlock", "2835": "univnet/modeling_univnet.py:UnivNetLvcBlock", "2836": "univnet/modeling_univnet.py:UnivNetModel", "2837": "fnet/modeling_fnet.py:_two_dim_matmul", "2838": "fnet/modeling_fnet.py:two_dim_matmul", "2839": "fnet/modeling_fnet.py:fftn", "2840": "fnet/modeling_fnet.py:FNetEmbeddings", "2841": "fnet/modeling_fnet.py:FNetBasicFourierTransform", "2842": "fnet/modeling_fnet.py:FNetBasicOutput", "2843": "fnet/modeling_fnet.py:FNetFourierTransform", "2844": "fnet/modeling_fnet.py:FNetIntermediate", "2845": "fnet/modeling_fnet.py:FNetOutput", "2846": "fnet/modeling_fnet.py:FNetLayer", "2847": "fnet/modeling_fnet.py:FNetEncoder", "2848": "fnet/modeling_fnet.py:FNetPooler", "2849": "fnet/modeling_fnet.py:FNetPredictionHeadTransform", "2850": "fnet/modeling_fnet.py:FNetLMPredictionHead", "2851": "fnet/modeling_fnet.py:FNetOnlyMLMHead", "2852": "fnet/modeling_fnet.py:FNetOnlyNSPHead", "2853": "fnet/modeling_fnet.py:FNetPreTrainingHeads", "2854": "fnet/modeling_fnet.py:FNetPreTrainedModel", "2855": "fnet/modeling_fnet.py:FNetForPreTrainingOutput", "2856": "fnet/modeling_fnet.py:FNetModel", "2857": "fnet/modeling_fnet.py:FNetForPreTraining", "2858": "fnet/modeling_fnet.py:FNetForMaskedLM", "2859": "fnet/modeling_fnet.py:FNetForNextSentencePrediction", "2860": "fnet/modeling_fnet.py:FNetForSequenceClassification", "2861": "fnet/modeling_fnet.py:FNetForMultipleChoice", "2862": "fnet/modeling_fnet.py:FNetForTokenClassification", "2863": "fnet/modeling_fnet.py:FNetForQuestionAnswering", "2864": "mobilenet_v1/modeling_mobilenet_v1.py:apply_tf_padding", "2865": "mobilenet_v1/modeling_mobilenet_v1.py:MobileNetV1ConvLayer", "2866": "mobilenet_v1/modeling_mobilenet_v1.py:MobileNetV1PreTrainedModel", "2867": "mobilenet_v1/modeling_mobilenet_v1.py:MobileNetV1Model", "2868": "mobilenet_v1/modeling_mobilenet_v1.py:MobileNetV1ForImageClassification", "2869": "jetmoe/modeling_jetmoe.py:load_balancing_loss_func", "2870": "jetmoe/modeling_jetmoe.py:JetMoeParallelExperts", "2871": "jetmoe/modeling_jetmoe.py:JetMoeTopKGating", "2872": "jetmoe/modeling_jetmoe.py:JetMoeMoE", "2873": "jetmoe/modeling_jetmoe.py:JetMoeMoA", "2874": "jetmoe/modeling_jetmoe.py:JetMoeRMSNorm", "2875": "jetmoe/modeling_jetmoe.py:JetMoeRotaryEmbedding", "2876": "jetmoe/modeling_jetmoe.py:rotate_half", "2877": "jetmoe/modeling_jetmoe.py:apply_rotary_pos_emb", "2878": "jetmoe/modeling_jetmoe.py:JetMoeAttention", "2879": "jetmoe/modeling_jetmoe.py:JetMoeSdpaAttention", "2880": "jetmoe/modeling_jetmoe.py:JetMoeFlashAttention2", "2881": "jetmoe/modeling_jetmoe.py:JetMoeBlock", "2882": "jetmoe/modeling_jetmoe.py:JetMoePreTrainedModel", "2883": "jetmoe/modeling_jetmoe.py:JetMoeModel", "2884": "jetmoe/modeling_jetmoe.py:JetMoeForCausalLM", "2885": "jetmoe/modeling_jetmoe.py:JetMoeForSequenceClassification", "2886": "dinov3_convnext/modeling_dinov3_convnext.py:drop_path", "2887": "dinov3_convnext/modeling_dinov3_convnext.py:DINOv3ConvNextDropPath", "2888": "dinov3_convnext/modeling_dinov3_convnext.py:DINOv3ConvNextLayerNorm", "2889": "dinov3_convnext/modeling_dinov3_convnext.py:DINOv3ConvNextLayer", "2890": "dinov3_convnext/modeling_dinov3_convnext.py:DINOv3ConvNextStage", "2891": "dinov3_convnext/modeling_dinov3_convnext.py:DINOv3ConvNextPreTrainedModel", "2892": "dinov3_convnext/modeling_dinov3_convnext.py:DINOv3ConvNextModel", "2893": "splinter/modeling_splinter.py:SplinterEmbeddings", "2894": "splinter/modeling_splinter.py:eager_attention_forward", "2895": "splinter/modeling_splinter.py:SplinterSelfAttention", "2896": "splinter/modeling_splinter.py:SplinterSelfOutput", "2897": "splinter/modeling_splinter.py:SplinterAttention", "2898": "splinter/modeling_splinter.py:SplinterIntermediate", "2899": "splinter/modeling_splinter.py:SplinterOutput", "2900": "splinter/modeling_splinter.py:SplinterLayer", "2901": "splinter/modeling_splinter.py:SplinterEncoder", "2902": "splinter/modeling_splinter.py:SplinterPreTrainedModel", "2903": "splinter/modeling_splinter.py:SplinterModel", "2904": "splinter/modeling_splinter.py:SplinterFullyConnectedLayer", "2905": "splinter/modeling_splinter.py:QuestionAwareSpanSelectionHead", "2906": "splinter/modeling_splinter.py:SplinterForQuestionAnswering", "2907": "splinter/modeling_splinter.py:SplinterForPreTrainingOutput", "2908": "splinter/modeling_splinter.py:SplinterForPreTraining", "2909": "vitpose/modeling_vitpose.py:VitPoseEstimatorOutput", "2910": "vitpose/modeling_vitpose.py:VitPosePreTrainedModel", "2911": "vitpose/modeling_vitpose.py:flip_back", "2912": "vitpose/modeling_vitpose.py:VitPoseSimpleDecoder", "2913": "vitpose/modeling_vitpose.py:VitPoseClassicDecoder", "2914": "vitpose/modeling_vitpose.py:VitPoseForPoseEstimation", "2915": "gpt2/modeling_gpt2.py:eager_attention_forward", "2916": "gpt2/modeling_gpt2.py:GPT2Attention", "2917": "gpt2/modeling_gpt2.py:GPT2MLP", "2918": "gpt2/modeling_gpt2.py:GPT2Block", "2919": "gpt2/modeling_gpt2.py:GPT2SequenceSummary", "2920": "gpt2/modeling_gpt2.py:GPT2PreTrainedModel", "2921": "gpt2/modeling_gpt2.py:GPT2DoubleHeadsModelOutput", "2922": "gpt2/modeling_gpt2.py:GPT2Model", "2923": "gpt2/modeling_gpt2.py:GPT2LMHeadModel", "2924": "gpt2/modeling_gpt2.py:GPT2DoubleHeadsModel", "2925": "gpt2/modeling_gpt2.py:GPT2ForSequenceClassification", "2926": "gpt2/modeling_gpt2.py:GPT2ForTokenClassification", "2927": "gpt2/modeling_gpt2.py:GPT2ForQuestionAnswering", "2928": "ibert/modeling_ibert.py:IBertEmbeddings", "2929": "ibert/modeling_ibert.py:IBertSelfAttention", "2930": "ibert/modeling_ibert.py:IBertSelfOutput", "2931": "ibert/modeling_ibert.py:IBertAttention", "2932": "ibert/modeling_ibert.py:IBertIntermediate", "2933": "ibert/modeling_ibert.py:IBertOutput", "2934": "ibert/modeling_ibert.py:IBertLayer", "2935": "ibert/modeling_ibert.py:IBertEncoder", "2936": "ibert/modeling_ibert.py:IBertPooler", "2937": "ibert/modeling_ibert.py:IBertPreTrainedModel", "2938": "ibert/modeling_ibert.py:IBertModel", "2939": "ibert/modeling_ibert.py:IBertForMaskedLM", "2940": "ibert/modeling_ibert.py:IBertLMHead", "2941": "ibert/modeling_ibert.py:IBertForSequenceClassification", "2942": "ibert/modeling_ibert.py:IBertForMultipleChoice", "2943": "ibert/modeling_ibert.py:IBertForTokenClassification", "2944": "ibert/modeling_ibert.py:IBertClassificationHead", "2945": "ibert/modeling_ibert.py:IBertForQuestionAnswering", "2946": "ibert/modeling_ibert.py:create_position_ids_from_input_ids", "2947": "depth_pro/modeling_depth_pro.py:DepthProOutput", "2948": "depth_pro/modeling_depth_pro.py:DepthProDepthEstimatorOutput", "2949": "depth_pro/modeling_depth_pro.py:split_to_patches", "2950": "depth_pro/modeling_depth_pro.py:reshape_features", "2951": "depth_pro/modeling_depth_pro.py:merge_patches", "2952": "depth_pro/modeling_depth_pro.py:reconstruct_feature_maps", "2953": "depth_pro/modeling_depth_pro.py:DepthProPatchEncoder", "2954": "depth_pro/modeling_depth_pro.py:DepthProImageEncoder", "2955": "depth_pro/modeling_depth_pro.py:DepthProEncoder", "2956": "depth_pro/modeling_depth_pro.py:DepthProFeatureUpsampleBlock", "2957": "depth_pro/modeling_depth_pro.py:DepthProFeatureUpsample", "2958": "depth_pro/modeling_depth_pro.py:DepthProFeatureProjection", "2959": "depth_pro/modeling_depth_pro.py:DepthProNeck", "2960": "depth_pro/modeling_depth_pro.py:DepthProPreTrainedModel", "2961": "depth_pro/modeling_depth_pro.py:DepthProModel", "2962": "depth_pro/modeling_depth_pro.py:DepthProPreActResidualLayer", "2963": "depth_pro/modeling_depth_pro.py:DepthProFeatureFusionLayer", "2964": "depth_pro/modeling_depth_pro.py:DepthProFeatureFusionStage", "2965": "depth_pro/modeling_depth_pro.py:DepthProFovEncoder", "2966": "depth_pro/modeling_depth_pro.py:DepthProFovHead", "2967": "depth_pro/modeling_depth_pro.py:DepthProFovModel", "2968": "depth_pro/modeling_depth_pro.py:DepthProDepthEstimationHead", "2969": "depth_pro/modeling_depth_pro.py:DepthProForDepthEstimation", "2970": "vitdet/modeling_vitdet.py:VitDetEmbeddings", "2971": "vitdet/modeling_vitdet.py:get_rel_pos", "2972": "vitdet/modeling_vitdet.py:add_decomposed_relative_positions", "2973": "vitdet/modeling_vitdet.py:VitDetAttention", "2974": "vitdet/modeling_vitdet.py:drop_path", "2975": "vitdet/modeling_vitdet.py:VitDetDropPath", "2976": "vitdet/modeling_vitdet.py:VitDetLayerNorm", "2977": "vitdet/modeling_vitdet.py:VitDetResBottleneckBlock", "2978": "vitdet/modeling_vitdet.py:VitDetMlp", "2979": "vitdet/modeling_vitdet.py:window_partition", "2980": "vitdet/modeling_vitdet.py:window_unpartition", "2981": "vitdet/modeling_vitdet.py:VitDetLayer", "2982": "vitdet/modeling_vitdet.py:VitDetEncoder", "2983": "vitdet/modeling_vitdet.py:caffe2_msra_fill", "2984": "vitdet/modeling_vitdet.py:VitDetPreTrainedModel", "2985": "vitdet/modeling_vitdet.py:VitDetModel", "2986": "vitdet/modeling_vitdet.py:VitDetBackbone", "2987": "textnet/modeling_textnet.py:TextNetConvLayer", "2988": "textnet/modeling_textnet.py:TextNetRepConvLayer", "2989": "textnet/modeling_textnet.py:TextNetStage", "2990": "textnet/modeling_textnet.py:TextNetEncoder", "2991": "textnet/modeling_textnet.py:TextNetPreTrainedModel", "2992": "textnet/modeling_textnet.py:TextNetModel", "2993": "textnet/modeling_textnet.py:TextNetForImageClassification", "2994": "textnet/modeling_textnet.py:TextNetBackbone", "2995": "gptj/modeling_gptj.py:create_sinusoidal_positions", "2996": "gptj/modeling_gptj.py:get_embed_positions", "2997": "gptj/modeling_gptj.py:rotate_every_two", "2998": "gptj/modeling_gptj.py:apply_rotary_pos_emb", "2999": "gptj/modeling_gptj.py:GPTJAttention", "3000": "gptj/modeling_gptj.py:GPTJFlashAttention2", "3001": "gptj/modeling_gptj.py:GPTJMLP", "3002": "gptj/modeling_gptj.py:GPTJBlock", "3003": "gptj/modeling_gptj.py:GPTJPreTrainedModel", "3004": "gptj/modeling_gptj.py:GPTJModel", "3005": "gptj/modeling_gptj.py:GPTJForCausalLM", "3006": "gptj/modeling_gptj.py:GPTJForSequenceClassification", "3007": "gptj/modeling_gptj.py:GPTJForQuestionAnswering", "3008": "xcodec/modeling_xcodec.py:XcodecOutput", "3009": "xcodec/modeling_xcodec.py:XcodecEncoderOutput", "3010": "xcodec/modeling_xcodec.py:XcodecDecoderOutput", "3011": "xcodec/modeling_xcodec.py:ResidualUnit", "3012": "xcodec/modeling_xcodec.py:SemanticEncoderBlock", "3013": "xcodec/modeling_xcodec.py:SemanticEncoder", "3014": "xcodec/modeling_xcodec.py:SemanticDecoderBlock", "3015": "xcodec/modeling_xcodec.py:SemanticDecoder", "3016": "xcodec/modeling_xcodec.py:XcodecEuclideanCodebook", "3017": "xcodec/modeling_xcodec.py:XcodecVectorQuantization", "3018": "xcodec/modeling_xcodec.py:XcodecResidualVectorQuantization", "3019": "xcodec/modeling_xcodec.py:XcodecPreTrainedModel", "3020": "xcodec/modeling_xcodec.py:XcodecModel", "3021": "udop/modeling_udop.py:BaseModelOutputWithAttentionMask", "3022": "udop/modeling_udop.py:get_visual_bbox", "3023": "udop/modeling_udop.py:pad_sequence", "3024": "udop/modeling_udop.py:combine_image_text_embeddings", "3025": "udop/modeling_udop.py:UdopPatchEmbeddings", "3026": "udop/modeling_udop.py:UdopPreTrainedModel", "3027": "udop/modeling_udop.py:UdopLayerNorm", "3028": "udop/modeling_udop.py:UdopDenseActDense", "3029": "udop/modeling_udop.py:UdopDenseGatedActDense", "3030": "udop/modeling_udop.py:UdopLayerFF", "3031": "udop/modeling_udop.py:UdopAttention", "3032": "udop/modeling_udop.py:UdopLayerSelfAttention", "3033": "udop/modeling_udop.py:UdopLayerCrossAttention", "3034": "udop/modeling_udop.py:UdopBlock", "3035": "udop/modeling_udop.py:UdopCellEmbeddings", "3036": "udop/modeling_udop.py:RelativePositionBiasBase", "3037": "udop/modeling_udop.py:RelativePositionBias1D", "3038": "udop/modeling_udop.py:RelativePositionBiasHorizontal", "3039": "udop/modeling_udop.py:RelativePositionBiasVertical", "3040": "udop/modeling_udop.py:RelativePositionBiasAggregated", "3041": "udop/modeling_udop.py:create_relative_bias", "3042": "udop/modeling_udop.py:UdopStack", "3043": "udop/modeling_udop.py:UdopModel", "3044": "udop/modeling_udop.py:UdopForConditionalGeneration", "3045": "udop/modeling_udop.py:UdopEncoderModel", "3046": "glm/modeling_glm.py:GlmMLP", "3047": "glm/modeling_glm.py:repeat_kv", "3048": "glm/modeling_glm.py:eager_attention_forward", "3049": "glm/modeling_glm.py:rotate_half", "3050": "glm/modeling_glm.py:apply_rotary_pos_emb", "3051": "glm/modeling_glm.py:GlmAttention", "3052": "glm/modeling_glm.py:GlmRMSNorm", "3053": "glm/modeling_glm.py:GlmRotaryEmbedding", "3054": "glm/modeling_glm.py:GlmDecoderLayer", "3055": "glm/modeling_glm.py:GlmPreTrainedModel", "3056": "glm/modeling_glm.py:GlmModel", "3057": "glm/modeling_glm.py:GlmForCausalLM", "3058": "glm/modeling_glm.py:GlmForSequenceClassification", "3059": "glm/modeling_glm.py:GlmForTokenClassification", "3060": "ctrl/modeling_ctrl.py:angle_defn", "3061": "ctrl/modeling_ctrl.py:positional_encoding", "3062": "ctrl/modeling_ctrl.py:scaled_dot_product_attention", "3063": "ctrl/modeling_ctrl.py:MultiHeadAttention", "3064": "ctrl/modeling_ctrl.py:point_wise_feed_forward_network", "3065": "ctrl/modeling_ctrl.py:EncoderLayer", "3066": "ctrl/modeling_ctrl.py:CTRLPreTrainedModel", "3067": "ctrl/modeling_ctrl.py:CTRLModel", "3068": "ctrl/modeling_ctrl.py:CTRLLMHeadModel", "3069": "ctrl/modeling_ctrl.py:CTRLForSequenceClassification", "3070": "llama/modeling_llama.py:LlamaRMSNorm", "3071": "llama/modeling_llama.py:LlamaRotaryEmbedding", "3072": "llama/modeling_llama.py:rotate_half", "3073": "llama/modeling_llama.py:apply_rotary_pos_emb", "3074": "llama/modeling_llama.py:LlamaMLP", "3075": "llama/modeling_llama.py:repeat_kv", "3076": "llama/modeling_llama.py:eager_attention_forward", "3077": "llama/modeling_llama.py:LlamaAttention", "3078": "llama/modeling_llama.py:LlamaDecoderLayer", "3079": "llama/modeling_llama.py:LlamaPreTrainedModel", "3080": "llama/modeling_llama.py:LlamaModel", "3081": "llama/modeling_llama.py:LlamaForCausalLM", "3082": "llama/modeling_llama.py:LlamaForSequenceClassification", "3083": "llama/modeling_llama.py:LlamaForQuestionAnswering", "3084": "llama/modeling_llama.py:LlamaForTokenClassification", "3085": "perceiver/modeling_perceiver.py:PerceiverModelOutput", "3086": "perceiver/modeling_perceiver.py:PerceiverDecoderOutput", "3087": "perceiver/modeling_perceiver.py:PerceiverMaskedLMOutput", "3088": "perceiver/modeling_perceiver.py:PerceiverClassifierOutput", "3089": "perceiver/modeling_perceiver.py:PerceiverEmbeddings", "3090": "perceiver/modeling_perceiver.py:PerceiverSelfAttention", "3091": "perceiver/modeling_perceiver.py:PerceiverSelfOutput", "3092": "perceiver/modeling_perceiver.py:PerceiverAttention", "3093": "perceiver/modeling_perceiver.py:PerceiverMLP", "3094": "perceiver/modeling_perceiver.py:PerceiverLayer", "3095": "perceiver/modeling_perceiver.py:PerceiverEncoder", "3096": "perceiver/modeling_perceiver.py:PerceiverPreTrainedModel", "3097": "perceiver/modeling_perceiver.py:PerceiverModel", "3098": "perceiver/modeling_perceiver.py:PerceiverForMaskedLM", "3099": "perceiver/modeling_perceiver.py:PerceiverForSequenceClassification", "3100": "perceiver/modeling_perceiver.py:PerceiverForImageClassificationLearned", "3101": "perceiver/modeling_perceiver.py:PerceiverForImageClassificationFourier", "3102": "perceiver/modeling_perceiver.py:PerceiverForImageClassificationConvProcessing", "3103": "perceiver/modeling_perceiver.py:PerceiverForOpticalFlow", "3104": "perceiver/modeling_perceiver.py:PerceiverForMultimodalAutoencoding", "3105": "perceiver/modeling_perceiver.py:build_position_encoding", "3106": "perceiver/modeling_perceiver.py:PerceiverAbstractDecoder", "3107": "perceiver/modeling_perceiver.py:PerceiverProjectionDecoder", "3108": "perceiver/modeling_perceiver.py:PerceiverBasicDecoder", "3109": "perceiver/modeling_perceiver.py:PerceiverClassificationDecoder", "3110": "perceiver/modeling_perceiver.py:PerceiverOpticalFlowDecoder", "3111": "perceiver/modeling_perceiver.py:PerceiverBasicVideoAutoencodingDecoder", "3112": "perceiver/modeling_perceiver.py:restructure", "3113": "perceiver/modeling_perceiver.py:PerceiverMultimodalDecoder", "3114": "perceiver/modeling_perceiver.py:space_to_depth", "3115": "perceiver/modeling_perceiver.py:Conv2dSamePadding", "3116": "perceiver/modeling_perceiver.py:Conv2DDownsample", "3117": "perceiver/modeling_perceiver.py:generate_fourier_features", "3118": "perceiver/modeling_perceiver.py:build_linear_positions", "3119": "perceiver/modeling_perceiver.py:PerceiverAbstractPositionEncoding", "3120": "perceiver/modeling_perceiver.py:PerceiverTrainablePositionEncoding", "3121": "perceiver/modeling_perceiver.py:_check_or_build_spatial_positions", "3122": "perceiver/modeling_perceiver.py:PerceiverFourierPositionEncoding", "3123": "perceiver/modeling_perceiver.py:AbstractPreprocessor", "3124": "perceiver/modeling_perceiver.py:PerceiverTextPreprocessor", "3125": "perceiver/modeling_perceiver.py:PerceiverEmbeddingDecoder", "3126": "perceiver/modeling_perceiver.py:PerceiverMultimodalPostprocessor", "3127": "perceiver/modeling_perceiver.py:PerceiverClassificationPostprocessor", "3128": "perceiver/modeling_perceiver.py:PerceiverAudioPostprocessor", "3129": "perceiver/modeling_perceiver.py:PerceiverProjectionPostprocessor", "3130": "perceiver/modeling_perceiver.py:PerceiverImagePreprocessor", "3131": "perceiver/modeling_perceiver.py:PerceiverOneHotPreprocessor", "3132": "perceiver/modeling_perceiver.py:PerceiverAudioPreprocessor", "3133": "perceiver/modeling_perceiver.py:PerceiverMultimodalPreprocessor", "3134": "dab_detr/modeling_dab_detr.py:DabDetrDecoderOutput", "3135": "dab_detr/modeling_dab_detr.py:DabDetrModelOutput", "3136": "dab_detr/modeling_dab_detr.py:DabDetrObjectDetectionOutput", "3137": "dab_detr/modeling_dab_detr.py:DabDetrFrozenBatchNorm2d", "3138": "dab_detr/modeling_dab_detr.py:replace_batch_norm", "3139": "dab_detr/modeling_dab_detr.py:DabDetrConvEncoder", "3140": "dab_detr/modeling_dab_detr.py:DabDetrConvModel", "3141": "dab_detr/modeling_dab_detr.py:DabDetrSinePositionEmbedding", "3142": "dab_detr/modeling_dab_detr.py:gen_sine_position_embeddings", "3143": "dab_detr/modeling_dab_detr.py:inverse_sigmoid", "3144": "dab_detr/modeling_dab_detr.py:DetrAttention", "3145": "dab_detr/modeling_dab_detr.py:DabDetrAttention", "3146": "dab_detr/modeling_dab_detr.py:DabDetrDecoderLayerSelfAttention", "3147": "dab_detr/modeling_dab_detr.py:DabDetrDecoderLayerCrossAttention", "3148": "dab_detr/modeling_dab_detr.py:DabDetrDecoderLayerFFN", "3149": "dab_detr/modeling_dab_detr.py:DabDetrEncoderLayer", "3150": "dab_detr/modeling_dab_detr.py:DabDetrDecoderLayer", "3151": "dab_detr/modeling_dab_detr.py:DabDetrMLP", "3152": "dab_detr/modeling_dab_detr.py:DabDetrPreTrainedModel", "3153": "dab_detr/modeling_dab_detr.py:DabDetrEncoder", "3154": "dab_detr/modeling_dab_detr.py:DabDetrDecoder", "3155": "dab_detr/modeling_dab_detr.py:DabDetrModel", "3156": "dab_detr/modeling_dab_detr.py:DabDetrMHAttentionMap", "3157": "dab_detr/modeling_dab_detr.py:DabDetrForObjectDetection", "3158": "reformer/modeling_reformer.py:ReformerDynamicCache", "3159": "reformer/modeling_reformer.py:_stable_argsort", "3160": "reformer/modeling_reformer.py:_get_least_common_mult_chunk_len", "3161": "reformer/modeling_reformer.py:_get_min_chunk_len", "3162": "reformer/modeling_reformer.py:AxialPositionEmbeddings", "3163": "reformer/modeling_reformer.py:PositionEmbeddings", "3164": "reformer/modeling_reformer.py:ReformerEmbeddings", "3165": "reformer/modeling_reformer.py:EfficientAttentionMixin", "3166": "reformer/modeling_reformer.py:LSHSelfAttention", "3167": "reformer/modeling_reformer.py:ReverseSort", "3168": "reformer/modeling_reformer.py:LocalSelfAttention", "3169": "reformer/modeling_reformer.py:ReformerSelfOutput", "3170": "reformer/modeling_reformer.py:ReformerAttention", "3171": "reformer/modeling_reformer.py:ReformerFeedForwardDense", "3172": "reformer/modeling_reformer.py:ReformerFeedForwardOutput", "3173": "reformer/modeling_reformer.py:ChunkReformerFeedForward", "3174": "reformer/modeling_reformer.py:ReformerLayer", "3175": "reformer/modeling_reformer.py:_ReversibleFunction", "3176": "reformer/modeling_reformer.py:ReformerEncoder", "3177": "reformer/modeling_reformer.py:ReformerOnlyLMHead", "3178": "reformer/modeling_reformer.py:ReformerPreTrainedModel", "3179": "reformer/modeling_reformer.py:ReformerModelOutput", "3180": "reformer/modeling_reformer.py:ReformerModelWithLMHeadOutput", "3181": "reformer/modeling_reformer.py:ReformerModel", "3182": "reformer/modeling_reformer.py:ReformerModelWithLMHead", "3183": "reformer/modeling_reformer.py:ReformerForMaskedLM", "3184": "reformer/modeling_reformer.py:ReformerForSequenceClassification", "3185": "reformer/modeling_reformer.py:ReformerClassificationHead", "3186": "reformer/modeling_reformer.py:ReformerForQuestionAnswering", "3187": "efficientloftr/modeling_efficientloftr.py:KeypointMatchingOutput", "3188": "efficientloftr/modeling_efficientloftr.py:compute_embeddings", "3189": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRRotaryEmbedding", "3190": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRConvNormLayer", "3191": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRRepVGGBlock", "3192": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRRepVGGStage", "3193": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRepVGG", "3194": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRAggregationLayer", "3195": "efficientloftr/modeling_efficientloftr.py:rotate_half", "3196": "efficientloftr/modeling_efficientloftr.py:apply_rotary_pos_emb", "3197": "efficientloftr/modeling_efficientloftr.py:repeat_kv", "3198": "efficientloftr/modeling_efficientloftr.py:eager_attention_forward", "3199": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRAttention", "3200": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRMLP", "3201": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRAggregatedAttention", "3202": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRLocalFeatureTransformerLayer", "3203": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRLocalFeatureTransformer", "3204": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTROutConvBlock", "3205": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRFineFusionLayer", "3206": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRPreTrainedModel", "3207": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRModel", "3208": "efficientloftr/modeling_efficientloftr.py:mask_border", "3209": "efficientloftr/modeling_efficientloftr.py:create_meshgrid", "3210": "efficientloftr/modeling_efficientloftr.py:spatial_expectation2d", "3211": "efficientloftr/modeling_efficientloftr.py:EfficientLoFTRForKeypointMatching", "3212": "timesfm/modeling_timesfm.py:TimesFmOutput", "3213": "timesfm/modeling_timesfm.py:TimesFmOutputForPrediction", "3214": "timesfm/modeling_timesfm.py:TimesFmMLP", "3215": "timesfm/modeling_timesfm.py:TimesFmResidualBlock", "3216": "timesfm/modeling_timesfm.py:TimesFmRMSNorm", "3217": "timesfm/modeling_timesfm.py:TimesFmPositionalEmbedding", "3218": "timesfm/modeling_timesfm.py:simple_eager_attention_forward", "3219": "timesfm/modeling_timesfm.py:TimesFmAttention", "3220": "timesfm/modeling_timesfm.py:TimesFmDecoderLayer", "3221": "timesfm/modeling_timesfm.py:TimesFmPreTrainedModel", "3222": "timesfm/modeling_timesfm.py:TimesFmModel", "3223": "timesfm/modeling_timesfm.py:TimesFmModelForPrediction", "3224": "depth_anything/modeling_depth_anything.py:DepthAnythingReassembleLayer", "3225": "depth_anything/modeling_depth_anything.py:DepthAnythingReassembleStage", "3226": "depth_anything/modeling_depth_anything.py:DepthAnythingPreActResidualLayer", "3227": "depth_anything/modeling_depth_anything.py:DepthAnythingFeatureFusionLayer", "3228": "depth_anything/modeling_depth_anything.py:DepthAnythingFeatureFusionStage", "3229": "depth_anything/modeling_depth_anything.py:DepthAnythingPreTrainedModel", "3230": "depth_anything/modeling_depth_anything.py:DepthAnythingNeck", "3231": "depth_anything/modeling_depth_anything.py:DepthAnythingDepthEstimationHead", "3232": "depth_anything/modeling_depth_anything.py:DepthAnythingForDepthEstimation", "3233": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeRMSNorm", "3234": "glm4v_moe/modeling_glm4v_moe.py:repeat_kv", "3235": "glm4v_moe/modeling_glm4v_moe.py:eager_attention_forward", "3236": "glm4v_moe/modeling_glm4v_moe.py:rotate_half", "3237": "glm4v_moe/modeling_glm4v_moe.py:apply_multimodal_rotary_pos_emb", "3238": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextAttention", "3239": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextTopkRouter", "3240": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextMoE", "3241": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextMLP", "3242": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextRMSNorm", "3243": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextDecoderLayer", "3244": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoePreTrainedModel", "3245": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeisionMlp", "3246": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeVisionPatchEmbed", "3247": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeVisionRotaryEmbedding", "3248": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeVisionPatchMerger", "3249": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeVisionEmbeddings", "3250": "glm4v_moe/modeling_glm4v_moe.py:apply_rotary_pos_emb_vision", "3251": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeVisionAttention", "3252": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeVisionBlock", "3253": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextRotaryEmbedding", "3254": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeModelOutputWithPast", "3255": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeVisionModel", "3256": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeTextModel", "3257": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeModel", "3258": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeCausalLMOutputWithPast", "3259": "glm4v_moe/modeling_glm4v_moe.py:Glm4vMoeForConditionalGeneration", "3260": "timm_backbone/modeling_timm_backbone.py:TimmBackbone", "3261": "dpt/modeling_dpt.py:BaseModelOutputWithIntermediateActivations", "3262": "dpt/modeling_dpt.py:BaseModelOutputWithPoolingAndIntermediateActivations", "3263": "dpt/modeling_dpt.py:DPTViTHybridEmbeddings", "3264": "dpt/modeling_dpt.py:DPTViTEmbeddings", "3265": "dpt/modeling_dpt.py:DPTViTPatchEmbeddings", "3266": "dpt/modeling_dpt.py:eager_attention_forward", "3267": "dpt/modeling_dpt.py:DPTSelfAttention", "3268": "dpt/modeling_dpt.py:DPTViTSelfOutput", "3269": "dpt/modeling_dpt.py:DPTViTAttention", "3270": "dpt/modeling_dpt.py:DPTViTIntermediate", "3271": "dpt/modeling_dpt.py:DPTViTOutput", "3272": "dpt/modeling_dpt.py:DPTViTLayer", "3273": "dpt/modeling_dpt.py:DPTViTEncoder", "3274": "dpt/modeling_dpt.py:DPTReassembleStage", "3275": "dpt/modeling_dpt.py:_get_backbone_hidden_size", "3276": "dpt/modeling_dpt.py:DPTReassembleLayer", "3277": "dpt/modeling_dpt.py:DPTFeatureFusionStage", "3278": "dpt/modeling_dpt.py:DPTPreActResidualLayer", "3279": "dpt/modeling_dpt.py:DPTFeatureFusionLayer", "3280": "dpt/modeling_dpt.py:DPTPreTrainedModel", "3281": "dpt/modeling_dpt.py:DPTModel", "3282": "dpt/modeling_dpt.py:DPTViTPooler", "3283": "dpt/modeling_dpt.py:DPTNeck", "3284": "dpt/modeling_dpt.py:DPTDepthEstimationHead", "3285": "dpt/modeling_dpt.py:DPTForDepthEstimation", "3286": "dpt/modeling_dpt.py:DPTSemanticSegmentationHead", "3287": "dpt/modeling_dpt.py:DPTAuxiliaryHead", "3288": "dpt/modeling_dpt.py:DPTForSemanticSegmentation", "3289": "gemma/modeling_gemma.py:GemmaRMSNorm", "3290": "gemma/modeling_gemma.py:GemmaMLP", "3291": "gemma/modeling_gemma.py:GemmaRotaryEmbedding", "3292": "gemma/modeling_gemma.py:rotate_half", "3293": "gemma/modeling_gemma.py:apply_rotary_pos_emb", "3294": "gemma/modeling_gemma.py:repeat_kv", "3295": "gemma/modeling_gemma.py:eager_attention_forward", "3296": "gemma/modeling_gemma.py:GemmaAttention", "3297": "gemma/modeling_gemma.py:GemmaDecoderLayer", "3298": "gemma/modeling_gemma.py:GemmaPreTrainedModel", "3299": "gemma/modeling_gemma.py:GemmaModel", "3300": "gemma/modeling_gemma.py:GemmaForCausalLM", "3301": "gemma/modeling_gemma.py:GemmaForSequenceClassification", "3302": "gemma/modeling_gemma.py:GemmaForTokenClassification", "3303": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextRMSNorm", "3304": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextFlexibleLinear", "3305": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextPreTrainedModel", "3306": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextConv1dPaddingCache", "3307": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextEmbeddings", "3308": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextLinear", "3309": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextRotaryEmbedding", "3310": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextGatingMLP", "3311": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:rotate_half", "3312": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:apply_rotary_pos_emb", "3313": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:repeat_kv", "3314": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextAttention", "3315": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextFlashAttention2", "3316": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextSdpaAttention", "3317": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextDecoderLayer", "3318": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextModel", "3319": "kyutai_speech_to_text/modeling_kyutai_speech_to_text.py:KyutaiSpeechToTextForConditionalGeneration", "3320": "metaclip_2/modeling_metaclip_2.py:MetaClip2TextEmbeddings", "3321": "metaclip_2/modeling_metaclip_2.py:MetaClip2VisionEmbeddings", "3322": "metaclip_2/modeling_metaclip_2.py:eager_attention_forward", "3323": "metaclip_2/modeling_metaclip_2.py:MetaClip2Attention", "3324": "metaclip_2/modeling_metaclip_2.py:MetaClip2MLP", "3325": "metaclip_2/modeling_metaclip_2.py:MetaClip2PreTrainedModel", "3326": "metaclip_2/modeling_metaclip_2.py:MetaClip2EncoderLayer", "3327": "metaclip_2/modeling_metaclip_2.py:MetaClip2Encoder", "3328": "metaclip_2/modeling_metaclip_2.py:MetaClip2TextTransformer", "3329": "metaclip_2/modeling_metaclip_2.py:MetaClip2TextModel", "3330": "metaclip_2/modeling_metaclip_2.py:MetaClip2TextModelOutput", "3331": "metaclip_2/modeling_metaclip_2.py:MetaClip2TextModelWithProjection", "3332": "metaclip_2/modeling_metaclip_2.py:MetaClip2Output", "3333": "metaclip_2/modeling_metaclip_2.py:contrastive_loss", "3334": "metaclip_2/modeling_metaclip_2.py:metaclip_2_loss", "3335": "metaclip_2/modeling_metaclip_2.py:_get_vector_norm", "3336": "metaclip_2/modeling_metaclip_2.py:MetaClip2Model", "3337": "metaclip_2/modeling_metaclip_2.py:MetaClip2VisionTransformer", "3338": "metaclip_2/modeling_metaclip_2.py:MetaClip2VisionModel", "3339": "metaclip_2/modeling_metaclip_2.py:MetaClip2VisionModelOutput", "3340": "metaclip_2/modeling_metaclip_2.py:MetaClip2VisionModelWithProjection", "3341": "metaclip_2/modeling_metaclip_2.py:MetaClip2ForImageClassification", "3342": "granite/modeling_granite.py:rotate_half", "3343": "granite/modeling_granite.py:apply_rotary_pos_emb", "3344": "granite/modeling_granite.py:repeat_kv", "3345": "granite/modeling_granite.py:eager_attention_forward", "3346": "granite/modeling_granite.py:GraniteAttention", "3347": "granite/modeling_granite.py:GraniteRMSNorm", "3348": "granite/modeling_granite.py:GraniteMLP", "3349": "granite/modeling_granite.py:GraniteDecoderLayer", "3350": "granite/modeling_granite.py:GranitePreTrainedModel", "3351": "granite/modeling_granite.py:GraniteRotaryEmbedding", "3352": "granite/modeling_granite.py:GraniteModel", "3353": "granite/modeling_granite.py:GraniteForCausalLM", "3354": "flava/modeling_flava.py:FlavaModelOutput", "3355": "flava/modeling_flava.py:FlavaLosses", "3356": "flava/modeling_flava.py:FlavaForPreTrainingOutput", "3357": "flava/modeling_flava.py:FlavaImageEmbeddings", "3358": "flava/modeling_flava.py:PatchEmbeddings", "3359": "flava/modeling_flava.py:FlavaTextEmbeddings", "3360": "flava/modeling_flava.py:FlavaSelfAttention", "3361": "flava/modeling_flava.py:FlavaSelfOutput", "3362": "flava/modeling_flava.py:FlavaAttention", "3363": "flava/modeling_flava.py:FlavaIntermediate", "3364": "flava/modeling_flava.py:FlavaOutput", "3365": "flava/modeling_flava.py:FlavaLayer", "3366": "flava/modeling_flava.py:FlavaEncoder", "3367": "flava/modeling_flava.py:FlavaPooler", "3368": "flava/modeling_flava.py:FlavaPreTrainedModel", "3369": "flava/modeling_flava.py:FlavaImageModel", "3370": "flava/modeling_flava.py:FlavaTextModel", "3371": "flava/modeling_flava.py:FlavaMultimodalModel", "3372": "flava/modeling_flava.py:FlavaModel", "3373": "flava/modeling_flava.py:FlavaImageCodebookResPath", "3374": "flava/modeling_flava.py:FlavaImageCodebookBlock", "3375": "flava/modeling_flava.py:FlavaImageCodebookLayerGroup", "3376": "flava/modeling_flava.py:FlavaImageCodebook", "3377": "flava/modeling_flava.py:FlavaPredictionHeadTransform", "3378": "flava/modeling_flava.py:FlavaMaskedPredictionHead", "3379": "flava/modeling_flava.py:FlavaITMHead", "3380": "flava/modeling_flava.py:FlavaGlobalContrastiveHead", "3381": "flava/modeling_flava.py:FlavaForPreTraining", "3382": "smolvlm/modeling_smolvlm.py:SmolVLMRMSNorm", "3383": "smolvlm/modeling_smolvlm.py:SmolVLMPreTrainedModel", "3384": "smolvlm/modeling_smolvlm.py:SmolVLMVisionEmbeddings", "3385": "smolvlm/modeling_smolvlm.py:eager_attention_forward", "3386": "smolvlm/modeling_smolvlm.py:SmolVLMVisionAttention", "3387": "smolvlm/modeling_smolvlm.py:SmolVLMVisionMLP", "3388": "smolvlm/modeling_smolvlm.py:SmolVLMEncoderLayer", "3389": "smolvlm/modeling_smolvlm.py:SmolVLMEncoder", "3390": "smolvlm/modeling_smolvlm.py:SmolVLMVisionTransformer", "3391": "smolvlm/modeling_smolvlm.py:SmolVLMBaseModelOutputWithPast", "3392": "smolvlm/modeling_smolvlm.py:SmolVLMSimpleMLP", "3393": "smolvlm/modeling_smolvlm.py:SmolVLMConnector", "3394": "smolvlm/modeling_smolvlm.py:SmolVLMModel", "3395": "smolvlm/modeling_smolvlm.py:SmolVLMCausalLMOutputWithPast", "3396": "smolvlm/modeling_smolvlm.py:SmolVLMForConditionalGeneration", "3397": "rembert/modeling_rembert.py:RemBertEmbeddings", "3398": "rembert/modeling_rembert.py:RemBertPooler", "3399": "rembert/modeling_rembert.py:RemBertSelfAttention", "3400": "rembert/modeling_rembert.py:RemBertSelfOutput", "3401": "rembert/modeling_rembert.py:RemBertAttention", "3402": "rembert/modeling_rembert.py:RemBertIntermediate", "3403": "rembert/modeling_rembert.py:RemBertOutput", "3404": "rembert/modeling_rembert.py:RemBertLayer", "3405": "rembert/modeling_rembert.py:RemBertEncoder", "3406": "rembert/modeling_rembert.py:RemBertPredictionHeadTransform", "3407": "rembert/modeling_rembert.py:RemBertLMPredictionHead", "3408": "rembert/modeling_rembert.py:RemBertOnlyMLMHead", "3409": "rembert/modeling_rembert.py:RemBertPreTrainedModel", "3410": "rembert/modeling_rembert.py:RemBertModel", "3411": "rembert/modeling_rembert.py:RemBertForMaskedLM", "3412": "rembert/modeling_rembert.py:RemBertForCausalLM", "3413": "rembert/modeling_rembert.py:RemBertForSequenceClassification", "3414": "rembert/modeling_rembert.py:RemBertForMultipleChoice", "3415": "rembert/modeling_rembert.py:RemBertForTokenClassification", "3416": "rembert/modeling_rembert.py:RemBertForQuestionAnswering", "3417": "granitemoeshared/modeling_granitemoeshared.py:GraniteFlashAttentionKwargs", "3418": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedMLP", "3419": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedRMSNorm", "3420": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedParallelExperts", "3421": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedTopKGating", "3422": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedMoE", "3423": "granitemoeshared/modeling_granitemoeshared.py:rotate_half", "3424": "granitemoeshared/modeling_granitemoeshared.py:apply_rotary_pos_emb", "3425": "granitemoeshared/modeling_granitemoeshared.py:repeat_kv", "3426": "granitemoeshared/modeling_granitemoeshared.py:eager_attention_forward", "3427": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedAttention", "3428": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedDecoderLayer", "3429": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedPreTrainedModel", "3430": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedRotaryEmbedding", "3431": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedModel", "3432": "granitemoeshared/modeling_granitemoeshared.py:load_balancing_loss_func", "3433": "granitemoeshared/modeling_granitemoeshared.py:GraniteMoeSharedForCausalLM", "3434": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyOutputWithPast", "3435": "musicgen_melody/modeling_musicgen_melody.py:shift_tokens_right", "3436": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodySinusoidalPositionalEmbedding", "3437": "musicgen_melody/modeling_musicgen_melody.py:eager_attention_forward", "3438": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyAttention", "3439": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyDecoderLayer", "3440": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyPreTrainedModel", "3441": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyDecoder", "3442": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyModel", "3443": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyForCausalLM", "3444": "musicgen_melody/modeling_musicgen_melody.py:MusicgenMelodyForConditionalGeneration", "3445": "cvt/modeling_cvt.py:BaseModelOutputWithCLSToken", "3446": "cvt/modeling_cvt.py:drop_path", "3447": "cvt/modeling_cvt.py:CvtDropPath", "3448": "cvt/modeling_cvt.py:CvtEmbeddings", "3449": "cvt/modeling_cvt.py:CvtConvEmbeddings", "3450": "cvt/modeling_cvt.py:CvtSelfAttentionConvProjection", "3451": "cvt/modeling_cvt.py:CvtSelfAttentionLinearProjection", "3452": "cvt/modeling_cvt.py:CvtSelfAttentionProjection", "3453": "cvt/modeling_cvt.py:CvtSelfAttention", "3454": "cvt/modeling_cvt.py:CvtSelfOutput", "3455": "cvt/modeling_cvt.py:CvtAttention", "3456": "cvt/modeling_cvt.py:CvtIntermediate", "3457": "cvt/modeling_cvt.py:CvtOutput", "3458": "cvt/modeling_cvt.py:CvtLayer", "3459": "cvt/modeling_cvt.py:CvtStage", "3460": "cvt/modeling_cvt.py:CvtEncoder", "3461": "cvt/modeling_cvt.py:CvtPreTrainedModel", "3462": "cvt/modeling_cvt.py:CvtModel", "3463": "cvt/modeling_cvt.py:CvtForImageClassification", "3464": "dinat/modeling_dinat.py:DinatEncoderOutput", "3465": "dinat/modeling_dinat.py:DinatModelOutput", "3466": "dinat/modeling_dinat.py:DinatImageClassifierOutput", "3467": "dinat/modeling_dinat.py:DinatEmbeddings", "3468": "dinat/modeling_dinat.py:DinatPatchEmbeddings", "3469": "dinat/modeling_dinat.py:DinatDownsampler", "3470": "dinat/modeling_dinat.py:drop_path", "3471": "dinat/modeling_dinat.py:DinatDropPath", "3472": "dinat/modeling_dinat.py:NeighborhoodAttention", "3473": "dinat/modeling_dinat.py:NeighborhoodAttentionOutput", "3474": "dinat/modeling_dinat.py:NeighborhoodAttentionModule", "3475": "dinat/modeling_dinat.py:DinatIntermediate", "3476": "dinat/modeling_dinat.py:DinatOutput", "3477": "dinat/modeling_dinat.py:DinatLayer", "3478": "dinat/modeling_dinat.py:DinatStage", "3479": "dinat/modeling_dinat.py:DinatEncoder", "3480": "dinat/modeling_dinat.py:DinatPreTrainedModel", "3481": "dinat/modeling_dinat.py:DinatModel", "3482": "dinat/modeling_dinat.py:DinatForImageClassification", "3483": "dinat/modeling_dinat.py:DinatBackbone", "3484": "moonshine/modeling_moonshine.py:MoonshineEncoderMLP", "3485": "moonshine/modeling_moonshine.py:MoonshineDecoderMLP", "3486": "moonshine/modeling_moonshine.py:repeat_kv", "3487": "moonshine/modeling_moonshine.py:eager_attention_forward", "3488": "moonshine/modeling_moonshine.py:rotate_half", "3489": "moonshine/modeling_moonshine.py:apply_rotary_pos_emb", "3490": "moonshine/modeling_moonshine.py:MoonshineAttention", "3491": "moonshine/modeling_moonshine.py:MoonshineRotaryEmbedding", "3492": "moonshine/modeling_moonshine.py:MoonshineEncoderLayer", "3493": "moonshine/modeling_moonshine.py:MoonshineDecoderLayer", "3494": "moonshine/modeling_moonshine.py:MoonshinePreTrainedModel", "3495": "moonshine/modeling_moonshine.py:MoonshineEncoder", "3496": "moonshine/modeling_moonshine.py:MoonshineDecoder", "3497": "moonshine/modeling_moonshine.py:_compute_mask_indices", "3498": "moonshine/modeling_moonshine.py:MoonshineModel", "3499": "moonshine/modeling_moonshine.py:shift_tokens_right", "3500": "moonshine/modeling_moonshine.py:MoonshineForConditionalGeneration", "3501": "aya_vision/modeling_aya_vision.py:AyaVisionMultiModalProjector", "3502": "aya_vision/modeling_aya_vision.py:AyaVisionPreTrainedModel", "3503": "aya_vision/modeling_aya_vision.py:AyaVisionCausalLMOutputWithPast", "3504": "aya_vision/modeling_aya_vision.py:AyaVisionModelOutputWithPast", "3505": "aya_vision/modeling_aya_vision.py:AyaVisionModel", "3506": "aya_vision/modeling_aya_vision.py:AyaVisionForConditionalGeneration", "3507": "detr/modeling_detr.py:DetrDecoderOutput", "3508": "detr/modeling_detr.py:DetrModelOutput", "3509": "detr/modeling_detr.py:DetrObjectDetectionOutput", "3510": "detr/modeling_detr.py:DetrSegmentationOutput", "3511": "detr/modeling_detr.py:DetrFrozenBatchNorm2d", "3512": "detr/modeling_detr.py:replace_batch_norm", "3513": "detr/modeling_detr.py:DetrConvEncoder", "3514": "detr/modeling_detr.py:DetrConvModel", "3515": "detr/modeling_detr.py:DetrSinePositionEmbedding", "3516": "detr/modeling_detr.py:DetrLearnedPositionEmbedding", "3517": "detr/modeling_detr.py:build_position_encoding", "3518": "detr/modeling_detr.py:DetrAttention", "3519": "detr/modeling_detr.py:DetrEncoderLayer", "3520": "detr/modeling_detr.py:DetrDecoderLayer", "3521": "detr/modeling_detr.py:DetrPreTrainedModel", "3522": "detr/modeling_detr.py:DetrEncoder", "3523": "detr/modeling_detr.py:DetrDecoder", "3524": "detr/modeling_detr.py:DetrModel", "3525": "detr/modeling_detr.py:DetrMLPPredictionHead", "3526": "detr/modeling_detr.py:DetrForObjectDetection", "3527": "detr/modeling_detr.py:DetrForSegmentation", "3528": "detr/modeling_detr.py:_expand", "3529": "detr/modeling_detr.py:DetrMaskHeadSmallConv", "3530": "detr/modeling_detr.py:DetrMHAttentionMap", "3531": "yoso/modeling_yoso.py:load_cuda_kernels", "3532": "yoso/modeling_yoso.py:to_contiguous", "3533": "yoso/modeling_yoso.py:normalize", "3534": "yoso/modeling_yoso.py:hashing", "3535": "yoso/modeling_yoso.py:YosoCumulation", "3536": "yoso/modeling_yoso.py:YosoLSHCumulation", "3537": "yoso/modeling_yoso.py:YosoEmbeddings", "3538": "yoso/modeling_yoso.py:YosoSelfAttention", "3539": "yoso/modeling_yoso.py:YosoSelfOutput", "3540": "yoso/modeling_yoso.py:YosoAttention", "3541": "yoso/modeling_yoso.py:YosoIntermediate", "3542": "yoso/modeling_yoso.py:YosoOutput", "3543": "yoso/modeling_yoso.py:YosoLayer", "3544": "yoso/modeling_yoso.py:YosoEncoder", "3545": "yoso/modeling_yoso.py:YosoPredictionHeadTransform", "3546": "yoso/modeling_yoso.py:YosoLMPredictionHead", "3547": "yoso/modeling_yoso.py:YosoOnlyMLMHead", "3548": "yoso/modeling_yoso.py:YosoPreTrainedModel", "3549": "yoso/modeling_yoso.py:YosoModel", "3550": "yoso/modeling_yoso.py:YosoForMaskedLM", "3551": "yoso/modeling_yoso.py:YosoClassificationHead", "3552": "yoso/modeling_yoso.py:YosoForSequenceClassification", "3553": "yoso/modeling_yoso.py:YosoForMultipleChoice", "3554": "yoso/modeling_yoso.py:YosoForTokenClassification", "3555": "yoso/modeling_yoso.py:YosoForQuestionAnswering", "3556": "dots1/modeling_dots1.py:Dots1RMSNorm", "3557": "dots1/modeling_dots1.py:Dots1RotaryEmbedding", "3558": "dots1/modeling_dots1.py:rotate_half", "3559": "dots1/modeling_dots1.py:apply_rotary_pos_emb", "3560": "dots1/modeling_dots1.py:repeat_kv", "3561": "dots1/modeling_dots1.py:eager_attention_forward", "3562": "dots1/modeling_dots1.py:Dots1Attention", "3563": "dots1/modeling_dots1.py:Dots1MLP", "3564": "dots1/modeling_dots1.py:Dots1MoE", "3565": "dots1/modeling_dots1.py:Dots1TopkRouter", "3566": "dots1/modeling_dots1.py:Dots1DecoderLayer", "3567": "dots1/modeling_dots1.py:Dots1PreTrainedModel", "3568": "dots1/modeling_dots1.py:Dots1Model", "3569": "dots1/modeling_dots1.py:Dots1ForCausalLM", "3570": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaRMSNorm", "3571": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaRotaryEmbedding", "3572": "recurrent_gemma/modeling_recurrent_gemma.py:rotate_half", "3573": "recurrent_gemma/modeling_recurrent_gemma.py:apply_rotary_pos_emb", "3574": "recurrent_gemma/modeling_recurrent_gemma.py:repeat_kv", "3575": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaSdpaAttention", "3576": "recurrent_gemma/modeling_recurrent_gemma.py:SqrtBoundDerivative", "3577": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaRglru", "3578": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaRecurrentBlock", "3579": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaMlp", "3580": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaDecoderLayer", "3581": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaPreTrainedModel", "3582": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaModel", "3583": "recurrent_gemma/modeling_recurrent_gemma.py:RecurrentGemmaForCausalLM", "3584": "chameleon/modeling_chameleon.py:ChameleonRMSNorm", "3585": "chameleon/modeling_chameleon.py:ChameleonRotaryEmbedding", "3586": "chameleon/modeling_chameleon.py:ChameleonLinearScalingRotaryEmbedding", "3587": "chameleon/modeling_chameleon.py:ChameleonDynamicNTKScalingRotaryEmbedding", "3588": "chameleon/modeling_chameleon.py:rotate_half", "3589": "chameleon/modeling_chameleon.py:apply_rotary_pos_emb", "3590": "chameleon/modeling_chameleon.py:ChameleonMLP", "3591": "chameleon/modeling_chameleon.py:ChameleonLayerNorm", "3592": "chameleon/modeling_chameleon.py:repeat_kv", "3593": "chameleon/modeling_chameleon.py:eager_attention_forward", "3594": "chameleon/modeling_chameleon.py:ChameleonAttention", "3595": "chameleon/modeling_chameleon.py:ChameleonDecoderLayer", "3596": "chameleon/modeling_chameleon.py:ChameleonSwinDecoderLayer", "3597": "chameleon/modeling_chameleon.py:ChameleonVQVAEVectorQuantizer", "3598": "chameleon/modeling_chameleon.py:ChameleonVQVAEEncoderConvDownsample", "3599": "chameleon/modeling_chameleon.py:ChameleonVQVAEEncoderResnetBlock", "3600": "chameleon/modeling_chameleon.py:ChameleonVQVAEEncoderAttnBlock", "3601": "chameleon/modeling_chameleon.py:ChameleonVQVAEEncoder", "3602": "chameleon/modeling_chameleon.py:ChameleonImageVocabularyMapping", "3603": "chameleon/modeling_chameleon.py:ChameleonPreTrainedModel", "3604": "chameleon/modeling_chameleon.py:ChameleonVQVAE", "3605": "chameleon/modeling_chameleon.py:ChameleonModel", "3606": "chameleon/modeling_chameleon.py:ChameleonForConditionalGeneration", "3607": "qwen3_next/modeling_qwen3_next.py:Qwen3NextRMSNormGated", "3608": "qwen3_next/modeling_qwen3_next.py:Qwen3NextDynamicCache", "3609": "qwen3_next/modeling_qwen3_next.py:Qwen3NextRotaryEmbedding", "3610": "qwen3_next/modeling_qwen3_next.py:Qwen3NextRMSNorm", "3611": "qwen3_next/modeling_qwen3_next.py:rotate_half", "3612": "qwen3_next/modeling_qwen3_next.py:apply_rotary_pos_emb", "3613": "qwen3_next/modeling_qwen3_next.py:repeat_kv", "3614": "qwen3_next/modeling_qwen3_next.py:eager_attention_forward", "3615": "qwen3_next/modeling_qwen3_next.py:Qwen3NextAttention", "3616": "qwen3_next/modeling_qwen3_next.py:apply_mask_to_padding_states", "3617": "qwen3_next/modeling_qwen3_next.py:torch_causal_conv1d_update", "3618": "qwen3_next/modeling_qwen3_next.py:l2norm", "3619": "qwen3_next/modeling_qwen3_next.py:torch_chunk_gated_delta_rule", "3620": "qwen3_next/modeling_qwen3_next.py:torch_recurrent_gated_delta_rule", "3621": "qwen3_next/modeling_qwen3_next.py:Qwen3NextGatedDeltaNet", "3622": "qwen3_next/modeling_qwen3_next.py:Qwen3NextMLP", "3623": "qwen3_next/modeling_qwen3_next.py:Qwen3NextSparseMoeBlock", "3624": "qwen3_next/modeling_qwen3_next.py:Qwen3NextDecoderLayer", "3625": "qwen3_next/modeling_qwen3_next.py:Qwen3NextPreTrainedModel", "3626": "qwen3_next/modeling_qwen3_next.py:Qwen3NextModel", "3627": "qwen3_next/modeling_qwen3_next.py:load_balancing_loss_func", "3628": "qwen3_next/modeling_qwen3_next.py:Qwen3NextForCausalLM", "3629": "qwen3_next/modeling_qwen3_next.py:Qwen3NextForSequenceClassification", "3630": "qwen3_next/modeling_qwen3_next.py:Qwen3NextForTokenClassification", "3631": "qwen3_next/modeling_qwen3_next.py:Qwen3NextForQuestionAnswering", "3632": "starcoder2/modeling_starcoder2.py:Starcoder2MLP", "3633": "starcoder2/modeling_starcoder2.py:rotate_half", "3634": "starcoder2/modeling_starcoder2.py:apply_rotary_pos_emb", "3635": "starcoder2/modeling_starcoder2.py:repeat_kv", "3636": "starcoder2/modeling_starcoder2.py:eager_attention_forward", "3637": "starcoder2/modeling_starcoder2.py:Starcoder2Attention", "3638": "starcoder2/modeling_starcoder2.py:Starcoder2DecoderLayer", "3639": "starcoder2/modeling_starcoder2.py:Starcoder2RotaryEmbedding", "3640": "starcoder2/modeling_starcoder2.py:Starcoder2PreTrainedModel", "3641": "starcoder2/modeling_starcoder2.py:Starcoder2Model", "3642": "starcoder2/modeling_starcoder2.py:Starcoder2ForCausalLM", "3643": "starcoder2/modeling_starcoder2.py:Starcoder2ForSequenceClassification", "3644": "starcoder2/modeling_starcoder2.py:Starcoder2ForTokenClassification", "3645": "sam_hq/modeling_sam_hq.py:SamHQVisionEncoderOutput", "3646": "sam_hq/modeling_sam_hq.py:SamHQMMaskDecoderOutputs", "3647": "sam_hq/modeling_sam_hq.py:SamHQImageSegmentationOutput", "3648": "sam_hq/modeling_sam_hq.py:SamHQVisionAttention", "3649": "sam_hq/modeling_sam_hq.py:SamHQMLPBlock", "3650": "sam_hq/modeling_sam_hq.py:SamHQVisionSdpaAttention", "3651": "sam_hq/modeling_sam_hq.py:SamHQVisionLayer", "3652": "sam_hq/modeling_sam_hq.py:SamHQPreTrainedModel", "3653": "sam_hq/modeling_sam_hq.py:SamHQPatchEmbeddings", "3654": "sam_hq/modeling_sam_hq.py:SamHQVisionNeck", "3655": "sam_hq/modeling_sam_hq.py:SamHQVisionEncoder", "3656": "sam_hq/modeling_sam_hq.py:SamHQLayerNorm", "3657": "sam_hq/modeling_sam_hq.py:eager_attention_forward", "3658": "sam_hq/modeling_sam_hq.py:SamHQAttention", "3659": "sam_hq/modeling_sam_hq.py:SamHQTwoWayAttentionBlock", "3660": "sam_hq/modeling_sam_hq.py:SamHQTwoWayTransformer", "3661": "sam_hq/modeling_sam_hq.py:SamHQFeedForward", "3662": "sam_hq/modeling_sam_hq.py:SamHQMaskDecoder", "3663": "sam_hq/modeling_sam_hq.py:SamHQVisionModel", "3664": "sam_hq/modeling_sam_hq.py:SamHQPositionalEmbedding", "3665": "sam_hq/modeling_sam_hq.py:SamHQMaskEmbedding", "3666": "sam_hq/modeling_sam_hq.py:SamHQPromptEncoder", "3667": "sam_hq/modeling_sam_hq.py:SamHQModel", "3668": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertRotaryPositionalEmbedding", "3669": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertRelPositionalEmbedding", "3670": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertFeatureProjection", "3671": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertFeedForward", "3672": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertConvolutionModule", "3673": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertSelfAttention", "3674": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertEncoderLayer", "3675": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertEncoder", "3676": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertAdapter", "3677": "wav2vec2_bert/modeling_wav2vec2_bert.py:_compute_new_attention_mask", "3678": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertAdapterLayer", "3679": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertPreTrainedModel", "3680": "wav2vec2_bert/modeling_wav2vec2_bert.py:_compute_mask_indices", "3681": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertModel", "3682": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertForCTC", "3683": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertForSequenceClassification", "3684": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertForAudioFrameClassification", "3685": "wav2vec2_bert/modeling_wav2vec2_bert.py:AMSoftmaxLoss", "3686": "wav2vec2_bert/modeling_wav2vec2_bert.py:TDNNLayer", "3687": "wav2vec2_bert/modeling_wav2vec2_bert.py:Wav2Vec2BertForXVector", "3688": "trocr/modeling_trocr.py:TrOCRLearnedPositionalEmbedding", "3689": "trocr/modeling_trocr.py:TrOCRScaledWordEmbedding", "3690": "trocr/modeling_trocr.py:TrOCRSinusoidalPositionalEmbedding", "3691": "trocr/modeling_trocr.py:TrOCRAttention", "3692": "trocr/modeling_trocr.py:TrOCRDecoderLayer", "3693": "trocr/modeling_trocr.py:TrOCRPreTrainedModel", "3694": "trocr/modeling_trocr.py:TrOCRDecoder", "3695": "trocr/modeling_trocr.py:TrOCRDecoderWrapper", "3696": "trocr/modeling_trocr.py:TrOCRForCausalLM", "3697": "florence2/modeling_florence2.py:drop_path", "3698": "florence2/modeling_florence2.py:Florence2VisionDropPath", "3699": "florence2/modeling_florence2.py:Florence2VisionLearnedAbsolutePositionEmbedding2D", "3700": "florence2/modeling_florence2.py:Florence2VisionPositionalEmbeddingCosine1D", "3701": "florence2/modeling_florence2.py:Florence2VisionMLP", "3702": "florence2/modeling_florence2.py:Florence2VisionConvEmbed", "3703": "florence2/modeling_florence2.py:eager_attention_forward", "3704": "florence2/modeling_florence2.py:Florence2VisionChannelAttention", "3705": "florence2/modeling_florence2.py:Florence2VisionChannelBlock", "3706": "florence2/modeling_florence2.py:Florence2VisionWindowAttention", "3707": "florence2/modeling_florence2.py:Florence2VisionSpatialBlock", "3708": "florence2/modeling_florence2.py:Florence2VisionBlock", "3709": "florence2/modeling_florence2.py:Florence2VisionPreTrainedModel", "3710": "florence2/modeling_florence2.py:Florence2VisionBackbone", "3711": "florence2/modeling_florence2.py:Florence2MultiModalProjector", "3712": "florence2/modeling_florence2.py:Florence2Seq2SeqModelOutput", "3713": "florence2/modeling_florence2.py:Florence2Seq2SeqLMOutput", "3714": "florence2/modeling_florence2.py:Florence2PreTrainedModel", "3715": "florence2/modeling_florence2.py:Florence2Model", "3716": "florence2/modeling_florence2.py:shift_tokens_right", "3717": "florence2/modeling_florence2.py:Florence2ForConditionalGeneration", "3718": "mixtral/modeling_mixtral.py:MixtralBlockSparseTop2MLP", "3719": "mixtral/modeling_mixtral.py:MixtralSparseMoeBlock", "3720": "mixtral/modeling_mixtral.py:MixtralRMSNorm", "3721": "mixtral/modeling_mixtral.py:rotate_half", "3722": "mixtral/modeling_mixtral.py:apply_rotary_pos_emb", "3723": "mixtral/modeling_mixtral.py:repeat_kv", "3724": "mixtral/modeling_mixtral.py:eager_attention_forward", "3725": "mixtral/modeling_mixtral.py:MixtralAttention", "3726": "mixtral/modeling_mixtral.py:MixtralDecoderLayer", "3727": "mixtral/modeling_mixtral.py:MixtralRotaryEmbedding", "3728": "mixtral/modeling_mixtral.py:MixtralPreTrainedModel", "3729": "mixtral/modeling_mixtral.py:MixtralModel", "3730": "mixtral/modeling_mixtral.py:load_balancing_loss_func", "3731": "mixtral/modeling_mixtral.py:MixtralForCausalLM", "3732": "mixtral/modeling_mixtral.py:MixtralForSequenceClassification", "3733": "mixtral/modeling_mixtral.py:MixtralForTokenClassification", "3734": "mixtral/modeling_mixtral.py:MixtralForQuestionAnswering", "3735": "kosmos2_5/modeling_kosmos2_5.py:_expand_mask", "3736": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5ModelOutput", "3737": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5ForConditionalGenerationModelOutput", "3738": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5LayerNorm", "3739": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5VisionEmbeddings", "3740": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5VisionMlp", "3741": "kosmos2_5/modeling_kosmos2_5.py:eager_attention_forward", "3742": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5VisionAttention", "3743": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5VisionLayer", "3744": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5VisionEncoder", "3745": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5TextSinusoidalPositionalEmbedding", "3746": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5TextFFN", "3747": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5TextAttention", "3748": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5TextBlock", "3749": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5TextTransformer", "3750": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5ImageToTextProjection", "3751": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5PreTrainedModel", "3752": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5VisionModel", "3753": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5TextModel", "3754": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5Model", "3755": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5TextForCausalLM", "3756": "kosmos2_5/modeling_kosmos2_5.py:Kosmos2_5ForConditionalGeneration", "3757": "qwen2_audio/modeling_qwen2_audio.py:Qwen2AudioCausalLMOutputWithPast", "3758": "qwen2_audio/modeling_qwen2_audio.py:eager_attention_forward", "3759": "qwen2_audio/modeling_qwen2_audio.py:Qwen2AudioAttention", "3760": "qwen2_audio/modeling_qwen2_audio.py:Qwen2AudioEncoderLayer", "3761": "qwen2_audio/modeling_qwen2_audio.py:Qwen2AudioPreTrainedModel", "3762": "qwen2_audio/modeling_qwen2_audio.py:Qwen2AudioEncoder", "3763": "qwen2_audio/modeling_qwen2_audio.py:Qwen2AudioMultiModalProjector", "3764": "qwen2_audio/modeling_qwen2_audio.py:Qwen2AudioForConditionalGeneration", "3765": "emu3/modeling_emu3.py:rotate_half", "3766": "emu3/modeling_emu3.py:apply_rotary_pos_emb", "3767": "emu3/modeling_emu3.py:repeat_kv", "3768": "emu3/modeling_emu3.py:eager_attention_forward", "3769": "emu3/modeling_emu3.py:Emu3Attention", "3770": "emu3/modeling_emu3.py:Emu3RMSNorm", "3771": "emu3/modeling_emu3.py:Emu3MLP", "3772": "emu3/modeling_emu3.py:Emu3DecoderLayer", "3773": "emu3/modeling_emu3.py:Emu3VQVAEVectorQuantizer", "3774": "emu3/modeling_emu3.py:Emu3VQVAEEncoderConvDownsample", "3775": "emu3/modeling_emu3.py:Emu3VQVAEEncoderConvUpsample", "3776": "emu3/modeling_emu3.py:Emu3VQVAEConv3d", "3777": "emu3/modeling_emu3.py:Emu3VQVAESpatialNorm", "3778": "emu3/modeling_emu3.py:Emu3VQVAETemporalUpsample", "3779": "emu3/modeling_emu3.py:Emu3VQVAETemporalDownsample", "3780": "emu3/modeling_emu3.py:Emu3VQVAETemporalResnetBlock", "3781": "emu3/modeling_emu3.py:Emu3VQVAEResnetBlock", "3782": "emu3/modeling_emu3.py:Emu3VQVAEAttentionBlock", "3783": "emu3/modeling_emu3.py:Emu3VQVAEGroupNorm", "3784": "emu3/modeling_emu3.py:Emu3VQVAEMiddleBlock", "3785": "emu3/modeling_emu3.py:Emu3VQVAEDownBlock", "3786": "emu3/modeling_emu3.py:Emu3VQVAEUpBlock", "3787": "emu3/modeling_emu3.py:Emu3VQVAEEncoder", "3788": "emu3/modeling_emu3.py:Emu3VQVAEDecoder", "3789": "emu3/modeling_emu3.py:Emu3VQVAE", "3790": "emu3/modeling_emu3.py:Emu3ImageVocabularyMapping", "3791": "emu3/modeling_emu3.py:Emu3PreTrainedModel", "3792": "emu3/modeling_emu3.py:Emu3RotaryEmbedding", "3793": "emu3/modeling_emu3.py:Emu3TextModel", "3794": "emu3/modeling_emu3.py:Emu3ForCausalLM", "3795": "emu3/modeling_emu3.py:Emu3Model", "3796": "emu3/modeling_emu3.py:Emu3ForConditionalGeneration", "3797": "colpali/modeling_colpali.py:ColPaliPreTrainedModel", "3798": "colpali/modeling_colpali.py:ColPaliForRetrievalOutput", "3799": "colpali/modeling_colpali.py:ColPaliForRetrieval", "3800": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionMLP", "3801": "phi4_multimodal/modeling_phi4_multimodal.py:simple_eager_attention_forward", "3802": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionAttention", "3803": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionEncoderLayer", "3804": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionEncoder", "3805": "phi4_multimodal/modeling_phi4_multimodal.py:_trunc_normal_", "3806": "phi4_multimodal/modeling_phi4_multimodal.py:trunc_normal_tf_", "3807": "phi4_multimodal/modeling_phi4_multimodal.py:variance_scaling_", "3808": "phi4_multimodal/modeling_phi4_multimodal.py:lecun_normal_", "3809": "phi4_multimodal/modeling_phi4_multimodal.py:default_flax_embed_init", "3810": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionPreTrainedModel", "3811": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionEmbeddings", "3812": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionMultiheadAttentionPoolingHead", "3813": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalVisionModel", "3814": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalImageEmbedding", "3815": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioMLP", "3816": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioAttention", "3817": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioDepthWiseSeparableConv1d", "3818": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioGluPointWiseConv", "3819": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioConvModule", "3820": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioConformerEncoderLayer", "3821": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioNemoConvSubsampling", "3822": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioRelativeAttentionBias", "3823": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioMeanVarianceNormLayer", "3824": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioPreTrainedModel", "3825": "phi4_multimodal/modeling_phi4_multimodal.py:unfold_tensor", "3826": "phi4_multimodal/modeling_phi4_multimodal.py:adaptive_enc_mask", "3827": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioModel", "3828": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAudioEmbedding", "3829": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalRMSNorm", "3830": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalMLP", "3831": "phi4_multimodal/modeling_phi4_multimodal.py:rotate_half", "3832": "phi4_multimodal/modeling_phi4_multimodal.py:repeat_kv", "3833": "phi4_multimodal/modeling_phi4_multimodal.py:eager_attention_forward", "3834": "phi4_multimodal/modeling_phi4_multimodal.py:apply_rotary_pos_emb", "3835": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalAttention", "3836": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalDecoderLayer", "3837": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalFeatureEmbedding", "3838": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalRotaryEmbedding", "3839": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalPreTrainedModel", "3840": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalModel", "3841": "phi4_multimodal/modeling_phi4_multimodal.py:Phi4MultimodalForCausalLM", "3842": "vitmatte/modeling_vitmatte.py:ImageMattingOutput", "3843": "vitmatte/modeling_vitmatte.py:VitMattePreTrainedModel", "3844": "vitmatte/modeling_vitmatte.py:VitMatteBasicConv3x3", "3845": "vitmatte/modeling_vitmatte.py:VitMatteConvStream", "3846": "vitmatte/modeling_vitmatte.py:VitMatteFusionBlock", "3847": "vitmatte/modeling_vitmatte.py:VitMatteHead", "3848": "vitmatte/modeling_vitmatte.py:VitMatteDetailCaptureModule", "3849": "vitmatte/modeling_vitmatte.py:VitMatteForImageMatting", "3850": "voxtral/modeling_voxtral.py:eager_attention_forward", "3851": "voxtral/modeling_voxtral.py:VoxtralAttention", "3852": "voxtral/modeling_voxtral.py:VoxtralEncoderLayer", "3853": "voxtral/modeling_voxtral.py:VoxtralPreTrainedModel", "3854": "voxtral/modeling_voxtral.py:VoxtralEncoder", "3855": "voxtral/modeling_voxtral.py:VoxtralMultiModalProjector", "3856": "voxtral/modeling_voxtral.py:VoxtralForConditionalGeneration", "3857": "deepseek_vl/modeling_deepseek_vl.py:DeepseekVLBaseModelOutputWithPast", "3858": "deepseek_vl/modeling_deepseek_vl.py:DeepseekVLCausalLMOutputWithPast", "3859": "deepseek_vl/modeling_deepseek_vl.py:DeepseekVLAligner", "3860": "deepseek_vl/modeling_deepseek_vl.py:DeepseekVLPreTrainedModel", "3861": "deepseek_vl/modeling_deepseek_vl.py:DeepseekVLModel", "3862": "deepseek_vl/modeling_deepseek_vl.py:DeepseekVLForConditionalGeneration", "3863": "marian/modeling_marian.py:shift_tokens_right", "3864": "marian/modeling_marian.py:MarianSinusoidalPositionalEmbedding", "3865": "marian/modeling_marian.py:eager_attention_forward", "3866": "marian/modeling_marian.py:MarianAttention", "3867": "marian/modeling_marian.py:MarianEncoderLayer", "3868": "marian/modeling_marian.py:MarianDecoderLayer", "3869": "marian/modeling_marian.py:MarianPreTrainedModel", "3870": "marian/modeling_marian.py:MarianEncoder", "3871": "marian/modeling_marian.py:MarianDecoder", "3872": "marian/modeling_marian.py:MarianModel", "3873": "marian/modeling_marian.py:MarianMTModel", "3874": "marian/modeling_marian.py:MarianDecoderWrapper", "3875": "marian/modeling_marian.py:MarianForCausalLM", "3876": "olmoe/modeling_olmoe.py:load_balancing_loss_func", "3877": "olmoe/modeling_olmoe.py:OlmoeRMSNorm", "3878": "olmoe/modeling_olmoe.py:OlmoeRotaryEmbedding", "3879": "olmoe/modeling_olmoe.py:rotate_half", "3880": "olmoe/modeling_olmoe.py:apply_rotary_pos_emb", "3881": "olmoe/modeling_olmoe.py:OlmoeMLP", "3882": "olmoe/modeling_olmoe.py:repeat_kv", "3883": "olmoe/modeling_olmoe.py:OlmoeAttention", "3884": "olmoe/modeling_olmoe.py:OlmoeFlashAttention2", "3885": "olmoe/modeling_olmoe.py:OlmoeSdpaAttention", "3886": "olmoe/modeling_olmoe.py:OlmoeSparseMoeBlock", "3887": "olmoe/modeling_olmoe.py:OlmoeDecoderLayer", "3888": "olmoe/modeling_olmoe.py:OlmoePreTrainedModel", "3889": "olmoe/modeling_olmoe.py:OlmoeModel", "3890": "olmoe/modeling_olmoe.py:OlmoeForCausalLM", "3891": "mimi/modeling_mimi.py:MimiOutput", "3892": "mimi/modeling_mimi.py:MimiConv1dPaddingCache", "3893": "mimi/modeling_mimi.py:MimiEncoderOutput", "3894": "mimi/modeling_mimi.py:MimiDecoderOutput", "3895": "mimi/modeling_mimi.py:MimiConv1d", "3896": "mimi/modeling_mimi.py:MimiConvTranspose1d", "3897": "mimi/modeling_mimi.py:MimiResnetBlock", "3898": "mimi/modeling_mimi.py:MimiEncoder", "3899": "mimi/modeling_mimi.py:MimiLayerScale", "3900": "mimi/modeling_mimi.py:MimiRotaryEmbedding", "3901": "mimi/modeling_mimi.py:rotate_half", "3902": "mimi/modeling_mimi.py:apply_rotary_pos_emb", "3903": "mimi/modeling_mimi.py:MimiMLP", "3904": "mimi/modeling_mimi.py:repeat_kv", "3905": "mimi/modeling_mimi.py:MimiAttention", "3906": "mimi/modeling_mimi.py:MimiFlashAttention2", "3907": "mimi/modeling_mimi.py:MimiSdpaAttention", "3908": "mimi/modeling_mimi.py:MimiTransformerLayer", "3909": "mimi/modeling_mimi.py:MimiTransformerModel", "3910": "mimi/modeling_mimi.py:MimiDecoder", "3911": "mimi/modeling_mimi.py:MimiEuclideanCodebook", "3912": "mimi/modeling_mimi.py:MimiVectorQuantization", "3913": "mimi/modeling_mimi.py:MimiResidualVectorQuantizer", "3914": "mimi/modeling_mimi.py:MimiSplitResidualVectorQuantizer", "3915": "mimi/modeling_mimi.py:MimiPreTrainedModel", "3916": "mimi/modeling_mimi.py:MimiModel", "3917": "altclip/modeling_altclip.py:contrastive_loss", "3918": "altclip/modeling_altclip.py:clip_loss", "3919": "altclip/modeling_altclip.py:AltCLIPOutput", "3920": "altclip/modeling_altclip.py:AltRobertaEmbeddings", "3921": "altclip/modeling_altclip.py:AltRobertaSelfAttention", "3922": "altclip/modeling_altclip.py:AltRobertaSelfOutput", "3923": "altclip/modeling_altclip.py:AltRobertaAttention", "3924": "altclip/modeling_altclip.py:AltRobertaIntermediate", "3925": "altclip/modeling_altclip.py:AltRobertaOutput", "3926": "altclip/modeling_altclip.py:AltRobertaLayer", "3927": "altclip/modeling_altclip.py:AltRobertaEncoder", "3928": "altclip/modeling_altclip.py:AltRobertaPooler", "3929": "altclip/modeling_altclip.py:eager_attention_forward", "3930": "altclip/modeling_altclip.py:AltCLIPAttention", "3931": "altclip/modeling_altclip.py:AltCLIPMLP", "3932": "altclip/modeling_altclip.py:AltCLIPEncoderLayer", "3933": "altclip/modeling_altclip.py:AltCLIPEncoder", "3934": "altclip/modeling_altclip.py:AltCLIPVisionEmbeddings", "3935": "altclip/modeling_altclip.py:AltCLIPPreTrainedModel", "3936": "altclip/modeling_altclip.py:AltCLIPVisionTransformer", "3937": "altclip/modeling_altclip.py:AltCLIPVisionModel", "3938": "altclip/modeling_altclip.py:AltRobertaModel", "3939": "altclip/modeling_altclip.py:AltCLIPTextModel", "3940": "altclip/modeling_altclip.py:AltCLIPModel", "3941": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionMLP", "3942": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionPatchEmbed", "3943": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionRotaryEmbedding", "3944": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionPatchMerger", "3945": "qwen3_vl/modeling_qwen3_vl.py:rotate_half", "3946": "qwen3_vl/modeling_qwen3_vl.py:apply_rotary_pos_emb_vision", "3947": "qwen3_vl/modeling_qwen3_vl.py:repeat_kv", "3948": "qwen3_vl/modeling_qwen3_vl.py:eager_attention_forward", "3949": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionAttention", "3950": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionBlock", "3951": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextRotaryEmbedding", "3952": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextRMSNorm", "3953": "qwen3_vl/modeling_qwen3_vl.py:apply_rotary_pos_emb", "3954": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextAttention", "3955": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextMLP", "3956": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextDecoderLayer", "3957": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLModelOutputWithPast", "3958": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLPreTrainedModel", "3959": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLVisionModel", "3960": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLTextModel", "3961": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLModel", "3962": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLCausalLMOutputWithPast", "3963": "qwen3_vl/modeling_qwen3_vl.py:Qwen3VLForConditionalGeneration", "3964": "glpn/modeling_glpn.py:drop_path", "3965": "glpn/modeling_glpn.py:GLPNDropPath", "3966": "glpn/modeling_glpn.py:GLPNOverlapPatchEmbeddings", "3967": "glpn/modeling_glpn.py:GLPNEfficientSelfAttention", "3968": "glpn/modeling_glpn.py:GLPNSelfOutput", "3969": "glpn/modeling_glpn.py:GLPNAttention", "3970": "glpn/modeling_glpn.py:GLPNDWConv", "3971": "glpn/modeling_glpn.py:GLPNMixFFN", "3972": "glpn/modeling_glpn.py:GLPNLayer", "3973": "glpn/modeling_glpn.py:GLPNEncoder", "3974": "glpn/modeling_glpn.py:GLPNPreTrainedModel", "3975": "glpn/modeling_glpn.py:GLPNModel", "3976": "glpn/modeling_glpn.py:GLPNSelectiveFeatureFusion", "3977": "glpn/modeling_glpn.py:GLPNDecoderStage", "3978": "glpn/modeling_glpn.py:GLPNDecoder", "3979": "glpn/modeling_glpn.py:SiLogLoss", "3980": "glpn/modeling_glpn.py:GLPNDepthEstimationHead", "3981": "glpn/modeling_glpn.py:GLPNForDepthEstimation", "3982": "superglue/modeling_superglue.py:concat_pairs", "3983": "superglue/modeling_superglue.py:normalize_keypoints", "3984": "superglue/modeling_superglue.py:log_sinkhorn_iterations", "3985": "superglue/modeling_superglue.py:log_optimal_transport", "3986": "superglue/modeling_superglue.py:arange_like", "3987": "superglue/modeling_superglue.py:KeypointMatchingOutput", "3988": "superglue/modeling_superglue.py:SuperGlueMultiLayerPerceptron", "3989": "superglue/modeling_superglue.py:SuperGlueKeypointEncoder", "3990": "superglue/modeling_superglue.py:SuperGlueSelfAttention", "3991": "superglue/modeling_superglue.py:SuperGlueSelfOutput", "3992": "superglue/modeling_superglue.py:SuperGlueAttention", "3993": "superglue/modeling_superglue.py:SuperGlueAttentionalPropagation", "3994": "superglue/modeling_superglue.py:SuperGlueAttentionalGNN", "3995": "superglue/modeling_superglue.py:SuperGlueFinalProjection", "3996": "superglue/modeling_superglue.py:SuperGluePreTrainedModel", "3997": "superglue/modeling_superglue.py:SuperGlueForKeypointMatching", "3998": "fsmt/modeling_fsmt.py:invert_mask", "3999": "fsmt/modeling_fsmt.py:triu_onnx", "4000": "fsmt/modeling_fsmt.py:_prepare_fsmt_decoder_inputs", "4001": "fsmt/modeling_fsmt.py:PretrainedFSMTModel", "4002": "fsmt/modeling_fsmt.py:_make_linear_from_emb", "4003": "fsmt/modeling_fsmt.py:_check_shapes", "4004": "fsmt/modeling_fsmt.py:shift_tokens_right", "4005": "fsmt/modeling_fsmt.py:make_padding_mask", "4006": "fsmt/modeling_fsmt.py:EncoderLayer", "4007": "fsmt/modeling_fsmt.py:FSMTEncoder", "4008": "fsmt/modeling_fsmt.py:DecoderLayer", "4009": "fsmt/modeling_fsmt.py:FSMTDecoder", "4010": "fsmt/modeling_fsmt.py:_reorder_buffer", "4011": "fsmt/modeling_fsmt.py:Attention", "4012": "fsmt/modeling_fsmt.py:fill_with_neg_inf", "4013": "fsmt/modeling_fsmt.py:_get_shape", "4014": "fsmt/modeling_fsmt.py:FSMTModel", "4015": "fsmt/modeling_fsmt.py:FSMTForConditionalGeneration", "4016": "fsmt/modeling_fsmt.py:SinusoidalPositionalEmbedding", "4017": "glm4/modeling_glm4.py:Glm4MLP", "4018": "glm4/modeling_glm4.py:Glm4DecoderLayer", "4019": "glm4/modeling_glm4.py:repeat_kv", "4020": "glm4/modeling_glm4.py:eager_attention_forward", "4021": "glm4/modeling_glm4.py:rotate_half", "4022": "glm4/modeling_glm4.py:apply_rotary_pos_emb", "4023": "glm4/modeling_glm4.py:Glm4Attention", "4024": "glm4/modeling_glm4.py:Glm4RMSNorm", "4025": "glm4/modeling_glm4.py:Glm4RotaryEmbedding", "4026": "glm4/modeling_glm4.py:Glm4PreTrainedModel", "4027": "glm4/modeling_glm4.py:Glm4Model", "4028": "glm4/modeling_glm4.py:Glm4ForCausalLM", "4029": "glm4/modeling_glm4.py:Glm4ForSequenceClassification", "4030": "glm4/modeling_glm4.py:Glm4ForTokenClassification", "4031": "owlvit/modeling_owlvit.py:contrastive_loss", "4032": "owlvit/modeling_owlvit.py:owlvit_loss", "4033": "owlvit/modeling_owlvit.py:OwlViTOutput", "4034": "owlvit/modeling_owlvit.py:_upcast", "4035": "owlvit/modeling_owlvit.py:box_area", "4036": "owlvit/modeling_owlvit.py:box_iou", "4037": "owlvit/modeling_owlvit.py:generalized_box_iou", "4038": "owlvit/modeling_owlvit.py:OwlViTObjectDetectionOutput", "4039": "owlvit/modeling_owlvit.py:OwlViTImageGuidedObjectDetectionOutput", "4040": "owlvit/modeling_owlvit.py:OwlViTVisionEmbeddings", "4041": "owlvit/modeling_owlvit.py:OwlViTTextEmbeddings", "4042": "owlvit/modeling_owlvit.py:OwlViTAttention", "4043": "owlvit/modeling_owlvit.py:OwlViTMLP", "4044": "owlvit/modeling_owlvit.py:OwlViTEncoderLayer", "4045": "owlvit/modeling_owlvit.py:OwlViTPreTrainedModel", "4046": "owlvit/modeling_owlvit.py:OwlViTEncoder", "4047": "owlvit/modeling_owlvit.py:OwlViTTextTransformer", "4048": "owlvit/modeling_owlvit.py:OwlViTTextModel", "4049": "owlvit/modeling_owlvit.py:OwlViTVisionTransformer", "4050": "owlvit/modeling_owlvit.py:OwlViTVisionModel", "4051": "owlvit/modeling_owlvit.py:OwlViTModel", "4052": "owlvit/modeling_owlvit.py:OwlViTBoxPredictionHead", "4053": "owlvit/modeling_owlvit.py:OwlViTClassPredictionHead", "4054": "owlvit/modeling_owlvit.py:OwlViTForObjectDetection", "4055": "llama4/modeling_llama4.py:Llama4TextExperts", "4056": "llama4/modeling_llama4.py:Llama4TextMLP", "4057": "llama4/modeling_llama4.py:Llama4TextL2Norm", "4058": "llama4/modeling_llama4.py:Llama4TextRMSNorm", "4059": "llama4/modeling_llama4.py:Llama4Router", "4060": "llama4/modeling_llama4.py:Llama4TextMoe", "4061": "llama4/modeling_llama4.py:Llama4TextRotaryEmbedding", "4062": "llama4/modeling_llama4.py:apply_rotary_emb", "4063": "llama4/modeling_llama4.py:repeat_kv", "4064": "llama4/modeling_llama4.py:eager_attention_forward", "4065": "llama4/modeling_llama4.py:vision_eager_attention_forward", "4066": "llama4/modeling_llama4.py:Llama4TextAttention", "4067": "llama4/modeling_llama4.py:Llama4TextDecoderLayer", "4068": "llama4/modeling_llama4.py:Llama4PreTrainedModel", "4069": "llama4/modeling_llama4.py:Llama4TextModel", "4070": "llama4/modeling_llama4.py:Llama4ForCausalLM", "4071": "llama4/modeling_llama4.py:Llama4CausalLMOutputWithPast", "4072": "llama4/modeling_llama4.py:Llama4VisionMLP2", "4073": "llama4/modeling_llama4.py:Llama4MultiModalProjector", "4074": "llama4/modeling_llama4.py:pixel_shuffle", "4075": "llama4/modeling_llama4.py:Llama4VisionPixelShuffleMLP", "4076": "llama4/modeling_llama4.py:reshape_for_broadcast", "4077": "llama4/modeling_llama4.py:vision_apply_rotary_emb", "4078": "llama4/modeling_llama4.py:Llama4VisionAttention", "4079": "llama4/modeling_llama4.py:Llama4VisionMLP", "4080": "llama4/modeling_llama4.py:Llama4VisionEncoderLayer", "4081": "llama4/modeling_llama4.py:Llama4VisionEncoder", "4082": "llama4/modeling_llama4.py:Llama4UnfoldConvolution", "4083": "llama4/modeling_llama4.py:Llama4VisionRotaryEmbedding", "4084": "llama4/modeling_llama4.py:Llama4VisionModel", "4085": "llama4/modeling_llama4.py:Llama4ForConditionalGeneration", "4086": "mamba/modeling_mamba.py:_lazy_load_causal_conv1d", "4087": "mamba/modeling_mamba.py:MambaCache", "4088": "mamba/modeling_mamba.py:MambaMixer", "4089": "mamba/modeling_mamba.py:MambaRMSNorm", "4090": "mamba/modeling_mamba.py:MambaBlock", "4091": "mamba/modeling_mamba.py:MambaPreTrainedModel", "4092": "mamba/modeling_mamba.py:MambaOutput", "4093": "mamba/modeling_mamba.py:MambaCausalLMOutput", "4094": "mamba/modeling_mamba.py:MambaModel", "4095": "mamba/modeling_mamba.py:MambaForCausalLM", "4096": "vision_encoder_decoder/modeling_vision_encoder_decoder.py:shift_tokens_right", "4097": "vision_encoder_decoder/modeling_vision_encoder_decoder.py:VisionEncoderDecoderModel", "4098": "t5gemma/modeling_t5gemma.py:T5GemmaRMSNorm", "4099": "t5gemma/modeling_t5gemma.py:T5GemmaMLP", "4100": "t5gemma/modeling_t5gemma.py:T5GemmaRotaryEmbedding", "4101": "t5gemma/modeling_t5gemma.py:rotate_half", "4102": "t5gemma/modeling_t5gemma.py:apply_rotary_pos_emb", "4103": "t5gemma/modeling_t5gemma.py:repeat_kv", "4104": "t5gemma/modeling_t5gemma.py:eager_attention_forward", "4105": "t5gemma/modeling_t5gemma.py:T5GemmaSelfAttention", "4106": "t5gemma/modeling_t5gemma.py:T5GemmaCrossAttention", "4107": "t5gemma/modeling_t5gemma.py:T5GemmaEncoderLayer", "4108": "t5gemma/modeling_t5gemma.py:T5GemmaDecoderLayer", "4109": "t5gemma/modeling_t5gemma.py:T5GemmaClassificationHead", "4110": "t5gemma/modeling_t5gemma.py:T5GemmaLMHead", "4111": "t5gemma/modeling_t5gemma.py:T5GemmaPreTrainedModel", "4112": "t5gemma/modeling_t5gemma.py:bidirectional_mask_function", "4113": "t5gemma/modeling_t5gemma.py:sliding_window_bidirectional_mask_function", "4114": "t5gemma/modeling_t5gemma.py:make_default_2d_attention_mask", "4115": "t5gemma/modeling_t5gemma.py:T5GemmaEncoder", "4116": "t5gemma/modeling_t5gemma.py:T5GemmaDecoder", "4117": "t5gemma/modeling_t5gemma.py:T5GemmaModel", "4118": "t5gemma/modeling_t5gemma.py:T5GemmaEncoderModel", "4119": "t5gemma/modeling_t5gemma.py:T5GemmaForConditionalGeneration", "4120": "t5gemma/modeling_t5gemma.py:T5GemmaForSequenceClassification", "4121": "t5gemma/modeling_t5gemma.py:T5GemmaForTokenClassification", "4122": "speech_encoder_decoder/modeling_speech_encoder_decoder.py:shift_tokens_right", "4123": "speech_encoder_decoder/modeling_speech_encoder_decoder.py:SpeechEncoderDecoderModel", "4124": "lightglue/modeling_lightglue.py:LightGlueKeypointMatchingOutput", "4125": "lightglue/modeling_lightglue.py:LightGluePositionalEncoder", "4126": "lightglue/modeling_lightglue.py:rotate_half", "4127": "lightglue/modeling_lightglue.py:apply_rotary_pos_emb", "4128": "lightglue/modeling_lightglue.py:repeat_kv", "4129": "lightglue/modeling_lightglue.py:eager_attention_forward", "4130": "lightglue/modeling_lightglue.py:LightGlueAttention", "4131": "lightglue/modeling_lightglue.py:LightGlueMLP", "4132": "lightglue/modeling_lightglue.py:LightGlueTransformerLayer", "4133": "lightglue/modeling_lightglue.py:sigmoid_log_double_softmax", "4134": "lightglue/modeling_lightglue.py:LightGlueMatchAssignmentLayer", "4135": "lightglue/modeling_lightglue.py:LightGlueTokenConfidenceLayer", "4136": "lightglue/modeling_lightglue.py:LightGluePreTrainedModel", "4137": "lightglue/modeling_lightglue.py:get_matches_from_scores", "4138": "lightglue/modeling_lightglue.py:normalize_keypoints", "4139": "lightglue/modeling_lightglue.py:LightGlueForKeypointMatching", "4140": "llava_next_video/modeling_llava_next_video.py:LlavaNextVideoModelOutputWithPast", "4141": "llava_next_video/modeling_llava_next_video.py:LlavaNextVideoCausalLMOutputWithPast", "4142": "llava_next_video/modeling_llava_next_video.py:LlavaNextVideoPooler", "4143": "llava_next_video/modeling_llava_next_video.py:LlavaNextVideoMultiModalProjector", "4144": "llava_next_video/modeling_llava_next_video.py:LlavaNextVideoPreTrainedModel", "4145": "llava_next_video/modeling_llava_next_video.py:get_anyres_image_grid_shape", "4146": "llava_next_video/modeling_llava_next_video.py:image_size_to_num_patches", "4147": "llava_next_video/modeling_llava_next_video.py:unpad_image", "4148": "llava_next_video/modeling_llava_next_video.py:LlavaNextVideoModel", "4149": "llava_next_video/modeling_llava_next_video.py:LlavaNextVideoForConditionalGeneration", "4150": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2GenerationOutput", "4151": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2TextToUnitDecoderOutput", "4152": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2TextToUnitOutput", "4153": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:shift_tokens_right", "4154": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:_compute_new_attention_mask", "4155": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:format_speech_generation_kwargs", "4156": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerFeatureProjection", "4157": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerFeedForward", "4158": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerConvolutionModule", "4159": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerSelfAttention", "4160": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerEncoderLayer", "4161": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerEncoder", "4162": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerAdapterLayer", "4163": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ConformerAdapter", "4164": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ScaledWordEmbedding", "4165": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2SinusoidalPositionalEmbedding", "4166": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2Attention", "4167": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2FeedForwardNetwork", "4168": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2EncoderLayer", "4169": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2DecoderLayer", "4170": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2TextToUnitDecoderLayer", "4171": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2PreTrainedModel", "4172": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2SpeechEncoder", "4173": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2Encoder", "4174": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2Decoder", "4175": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2TextToUnitDecoder", "4176": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2TextToUnitModel", "4177": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2TextToUnitForConditionalGeneration", "4178": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:HifiGanResidualBlock", "4179": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2VariancePredictor", "4180": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2HifiGan", "4181": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2CodeHifiGan", "4182": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ForTextToText", "4183": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ForSpeechToText", "4184": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ForTextToSpeech", "4185": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2ForSpeechToSpeech", "4186": "seamless_m4t_v2/modeling_seamless_m4t_v2.py:SeamlessM4Tv2Model", "4187": "convnext/modeling_convnext.py:drop_path", "4188": "convnext/modeling_convnext.py:ConvNextDropPath", "4189": "convnext/modeling_convnext.py:ConvNextLayerNorm", "4190": "convnext/modeling_convnext.py:ConvNextEmbeddings", "4191": "convnext/modeling_convnext.py:ConvNextLayer", "4192": "convnext/modeling_convnext.py:ConvNextStage", "4193": "convnext/modeling_convnext.py:ConvNextEncoder", "4194": "convnext/modeling_convnext.py:ConvNextPreTrainedModel", "4195": "convnext/modeling_convnext.py:ConvNextModel", "4196": "convnext/modeling_convnext.py:ConvNextForImageClassification", "4197": "convnext/modeling_convnext.py:ConvNextBackbone", "4198": "oneformer/modeling_oneformer.py:_get_clones", "4199": "oneformer/modeling_oneformer.py:multi_scale_deformable_attention", "4200": "oneformer/modeling_oneformer.py:dice_loss", "4201": "oneformer/modeling_oneformer.py:sigmoid_cross_entropy_loss", "4202": "oneformer/modeling_oneformer.py:pair_wise_dice_loss", "4203": "oneformer/modeling_oneformer.py:pair_wise_sigmoid_cross_entropy_loss", "4204": "oneformer/modeling_oneformer.py:sample_point", "4205": "oneformer/modeling_oneformer.py:OneFormerHungarianMatcher", "4206": "oneformer/modeling_oneformer.py:OneFormerLoss", "4207": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderOutput", "4208": "oneformer/modeling_oneformer.py:OneFormerPixelDecoderOutput", "4209": "oneformer/modeling_oneformer.py:OneFormerPixelLevelModuleOutput", "4210": "oneformer/modeling_oneformer.py:OneFormerModelOutput", "4211": "oneformer/modeling_oneformer.py:OneFormerForUniversalSegmentationOutput", "4212": "oneformer/modeling_oneformer.py:OneFormerPixelDecoderFrozenBatchNorm2d", "4213": "oneformer/modeling_oneformer.py:OneFormerPixelDecoderEncoderMultiscaleDeformableAttention", "4214": "oneformer/modeling_oneformer.py:OneFormerPixelDecoderEncoderLayer", "4215": "oneformer/modeling_oneformer.py:OneFormerPixelDecoderEncoderOnly", "4216": "oneformer/modeling_oneformer.py:OneFormerPixelDecoder", "4217": "oneformer/modeling_oneformer.py:OneFormerPixelLevelModule", "4218": "oneformer/modeling_oneformer.py:OneFormerAttention", "4219": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderSelfAttentionLayer", "4220": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderCrossAttentionLayer", "4221": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderFFNLayer", "4222": "oneformer/modeling_oneformer.py:OneFormerMLPPredictionHead", "4223": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderLayer", "4224": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderQueryTransformerDecoder", "4225": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderQueryTransformerDecoderLayer", "4226": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoderQueryTransformer", "4227": "oneformer/modeling_oneformer.py:OneFormerTransformerDecoder", "4228": "oneformer/modeling_oneformer.py:OneFormerTransformerModule", "4229": "oneformer/modeling_oneformer.py:OneFormerSinePositionEmbedding", "4230": "oneformer/modeling_oneformer.py:PredictionBlock", "4231": "oneformer/modeling_oneformer.py:OneFormerTextMapperAttention", "4232": "oneformer/modeling_oneformer.py:OneFormerTextTransformerDecoderLayer", "4233": "oneformer/modeling_oneformer.py:OneFormerTextContextDecoder", "4234": "oneformer/modeling_oneformer.py:OneFormerTextMLP", "4235": "oneformer/modeling_oneformer.py:OneFormerTextTransformerLayer", "4236": "oneformer/modeling_oneformer.py:OneFormerTextTransformer", "4237": "oneformer/modeling_oneformer.py:OneFormerTextEncoder", "4238": "oneformer/modeling_oneformer.py:OneFormerTextMapper", "4239": "oneformer/modeling_oneformer.py:OneFormerTaskModel", "4240": "oneformer/modeling_oneformer.py:OneFormerPreTrainedModel", "4241": "oneformer/modeling_oneformer.py:OneFormerModel", "4242": "oneformer/modeling_oneformer.py:OneFormerForUniversalSegmentation", "4243": "efficientnet/modeling_efficientnet.py:round_filters", "4244": "efficientnet/modeling_efficientnet.py:correct_pad", "4245": "efficientnet/modeling_efficientnet.py:EfficientNetEmbeddings", "4246": "efficientnet/modeling_efficientnet.py:EfficientNetDepthwiseConv2d", "4247": "efficientnet/modeling_efficientnet.py:EfficientNetExpansionLayer", "4248": "efficientnet/modeling_efficientnet.py:EfficientNetDepthwiseLayer", "4249": "efficientnet/modeling_efficientnet.py:EfficientNetSqueezeExciteLayer", "4250": "efficientnet/modeling_efficientnet.py:EfficientNetFinalBlockLayer", "4251": "efficientnet/modeling_efficientnet.py:EfficientNetBlock", "4252": "efficientnet/modeling_efficientnet.py:EfficientNetEncoder", "4253": "efficientnet/modeling_efficientnet.py:EfficientNetPreTrainedModel", "4254": "efficientnet/modeling_efficientnet.py:EfficientNetModel", "4255": "efficientnet/modeling_efficientnet.py:EfficientNetForImageClassification", "4256": "mobilebert/modeling_mobilebert.py:NoNorm", "4257": "mobilebert/modeling_mobilebert.py:MobileBertEmbeddings", "4258": "mobilebert/modeling_mobilebert.py:eager_attention_forward", "4259": "mobilebert/modeling_mobilebert.py:MobileBertSelfAttention", "4260": "mobilebert/modeling_mobilebert.py:MobileBertSelfOutput", "4261": "mobilebert/modeling_mobilebert.py:MobileBertAttention", "4262": "mobilebert/modeling_mobilebert.py:MobileBertIntermediate", "4263": "mobilebert/modeling_mobilebert.py:OutputBottleneck", "4264": "mobilebert/modeling_mobilebert.py:MobileBertOutput", "4265": "mobilebert/modeling_mobilebert.py:BottleneckLayer", "4266": "mobilebert/modeling_mobilebert.py:Bottleneck", "4267": "mobilebert/modeling_mobilebert.py:FFNOutput", "4268": "mobilebert/modeling_mobilebert.py:FFNLayer", "4269": "mobilebert/modeling_mobilebert.py:MobileBertLayer", "4270": "mobilebert/modeling_mobilebert.py:MobileBertEncoder", "4271": "mobilebert/modeling_mobilebert.py:MobileBertPooler", "4272": "mobilebert/modeling_mobilebert.py:MobileBertPredictionHeadTransform", "4273": "mobilebert/modeling_mobilebert.py:MobileBertLMPredictionHead", "4274": "mobilebert/modeling_mobilebert.py:MobileBertOnlyMLMHead", "4275": "mobilebert/modeling_mobilebert.py:MobileBertPreTrainingHeads", "4276": "mobilebert/modeling_mobilebert.py:MobileBertPreTrainedModel", "4277": "mobilebert/modeling_mobilebert.py:MobileBertForPreTrainingOutput", "4278": "mobilebert/modeling_mobilebert.py:MobileBertModel", "4279": "mobilebert/modeling_mobilebert.py:MobileBertForPreTraining", "4280": "mobilebert/modeling_mobilebert.py:MobileBertForMaskedLM", "4281": "mobilebert/modeling_mobilebert.py:MobileBertOnlyNSPHead", "4282": "mobilebert/modeling_mobilebert.py:MobileBertForNextSentencePrediction", "4283": "mobilebert/modeling_mobilebert.py:MobileBertForSequenceClassification", "4284": "mobilebert/modeling_mobilebert.py:MobileBertForQuestionAnswering", "4285": "mobilebert/modeling_mobilebert.py:MobileBertForMultipleChoice", "4286": "mobilebert/modeling_mobilebert.py:MobileBertForTokenClassification", "4287": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2PreTrainedModel", "4288": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2LearnableAffineBlock", "4289": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2ConvLayer", "4290": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2ConvLayerLight", "4291": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2Embeddings", "4292": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2BasicLayer", "4293": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2Stage", "4294": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2Encoder", "4295": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2Backbone", "4296": "hgnet_v2/modeling_hgnet_v2.py:HGNetV2ForImageClassification", "4297": "sam/modeling_sam.py:SamVisionEncoderOutput", "4298": "sam/modeling_sam.py:SamImageSegmentationOutput", "4299": "sam/modeling_sam.py:SamPatchEmbeddings", "4300": "sam/modeling_sam.py:SamMLPBlock", "4301": "sam/modeling_sam.py:SamLayerNorm", "4302": "sam/modeling_sam.py:eager_attention_forward", "4303": "sam/modeling_sam.py:SamAttention", "4304": "sam/modeling_sam.py:SamTwoWayAttentionBlock", "4305": "sam/modeling_sam.py:SamTwoWayTransformer", "4306": "sam/modeling_sam.py:SamFeedForward", "4307": "sam/modeling_sam.py:SamMaskDecoder", "4308": "sam/modeling_sam.py:SamPositionalEmbedding", "4309": "sam/modeling_sam.py:SamMaskEmbedding", "4310": "sam/modeling_sam.py:SamPromptEncoder", "4311": "sam/modeling_sam.py:SamVisionAttention", "4312": "sam/modeling_sam.py:SamVisionSdpaAttention", "4313": "sam/modeling_sam.py:SamVisionLayer", "4314": "sam/modeling_sam.py:SamVisionNeck", "4315": "sam/modeling_sam.py:SamPreTrainedModel", "4316": "sam/modeling_sam.py:SamVisionEncoder", "4317": "sam/modeling_sam.py:SamVisionModel", "4318": "sam/modeling_sam.py:SamModel", "4319": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLHybridBaseModelOutputWithPast", "4320": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLHybridCausalLMOutputWithPast", "4321": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLHybridLayerNorm", "4322": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLSamVisionNeck", "4323": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLSamVisionProj", "4324": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLHybridAligner", "4325": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLHybridPreTrainedModel", "4326": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLHybridModel", "4327": "deepseek_vl_hybrid/modeling_deepseek_vl_hybrid.py:DeepseekVLHybridForConditionalGeneration", "4328": "markuplm/modeling_markuplm.py:XPathEmbeddings", "4329": "markuplm/modeling_markuplm.py:MarkupLMEmbeddings", "4330": "markuplm/modeling_markuplm.py:MarkupLMSelfOutput", "4331": "markuplm/modeling_markuplm.py:MarkupLMIntermediate", "4332": "markuplm/modeling_markuplm.py:MarkupLMOutput", "4333": "markuplm/modeling_markuplm.py:MarkupLMPooler", "4334": "markuplm/modeling_markuplm.py:MarkupLMPredictionHeadTransform", "4335": "markuplm/modeling_markuplm.py:MarkupLMLMPredictionHead", "4336": "markuplm/modeling_markuplm.py:MarkupLMOnlyMLMHead", "4337": "markuplm/modeling_markuplm.py:eager_attention_forward", "4338": "markuplm/modeling_markuplm.py:MarkupLMSelfAttention", "4339": "markuplm/modeling_markuplm.py:MarkupLMAttention", "4340": "markuplm/modeling_markuplm.py:MarkupLMLayer", "4341": "markuplm/modeling_markuplm.py:MarkupLMEncoder", "4342": "markuplm/modeling_markuplm.py:MarkupLMPreTrainedModel", "4343": "markuplm/modeling_markuplm.py:MarkupLMModel", "4344": "markuplm/modeling_markuplm.py:MarkupLMForQuestionAnswering", "4345": "markuplm/modeling_markuplm.py:MarkupLMForTokenClassification", "4346": "markuplm/modeling_markuplm.py:MarkupLMForSequenceClassification", "4347": "data2vec/modeling_data2vec_vision.py:Data2VecVisionModelOutputWithPooling", "4348": "data2vec/modeling_data2vec_vision.py:drop_path", "4349": "data2vec/modeling_data2vec_vision.py:Data2VecVisionDropPath", "4350": "data2vec/modeling_data2vec_vision.py:Data2VecVisionEmbeddings", "4351": "data2vec/modeling_data2vec_vision.py:Data2VecVisionPatchEmbeddings", "4352": "data2vec/modeling_data2vec_vision.py:Data2VecVisionSelfAttention", "4353": "data2vec/modeling_data2vec_vision.py:Data2VecVisionSdpaSelfAttention", "4354": "data2vec/modeling_data2vec_vision.py:Data2VecVisionSelfOutput", "4355": "data2vec/modeling_data2vec_vision.py:Data2VecVisionAttention", "4356": "data2vec/modeling_data2vec_vision.py:Data2VecVisionIntermediate", "4357": "data2vec/modeling_data2vec_vision.py:Data2VecVisionOutput", "4358": "data2vec/modeling_data2vec_vision.py:Data2VecVisionLayer", "4359": "data2vec/modeling_data2vec_vision.py:Data2VecVisionRelativePositionBias", "4360": "data2vec/modeling_data2vec_vision.py:Data2VecVisionEncoder", "4361": "data2vec/modeling_data2vec_vision.py:Data2VecVisionPreTrainedModel", "4362": "data2vec/modeling_data2vec_vision.py:Data2VecVisionModel", "4363": "data2vec/modeling_data2vec_vision.py:Data2VecVisionPooler", "4364": "data2vec/modeling_data2vec_vision.py:Data2VecVisionForImageClassification", "4365": "data2vec/modeling_data2vec_vision.py:Data2VecVisionConvModule", "4366": "data2vec/modeling_data2vec_vision.py:Data2VecVisionPyramidPoolingBlock", "4367": "data2vec/modeling_data2vec_vision.py:Data2VecVisionPyramidPoolingModule", "4368": "data2vec/modeling_data2vec_vision.py:Data2VecVisionUperHead", "4369": "data2vec/modeling_data2vec_vision.py:Data2VecVisionFCNHead", "4370": "data2vec/modeling_data2vec_vision.py:Data2VecVisionForSemanticSegmentation", "4371": "data2vec/modeling_data2vec_audio.py:Data2VecAudioConvLayer", "4372": "data2vec/modeling_data2vec_audio.py:Data2VecAudioPadLayer", "4373": "data2vec/modeling_data2vec_audio.py:Data2VecAudioPositionalConvLayer", "4374": "data2vec/modeling_data2vec_audio.py:Data2VecAudioPositionalConvEmbedding", "4375": "data2vec/modeling_data2vec_audio.py:Data2VecAudioFeatureEncoder", "4376": "data2vec/modeling_data2vec_audio.py:Data2VecAudioFeatureProjection", "4377": "data2vec/modeling_data2vec_audio.py:eager_attention_forward", "4378": "data2vec/modeling_data2vec_audio.py:Data2VecAudioAttention", "4379": "data2vec/modeling_data2vec_audio.py:Data2VecAudioFeedForward", "4380": "data2vec/modeling_data2vec_audio.py:Data2VecAudioEncoderLayer", "4381": "data2vec/modeling_data2vec_audio.py:Data2VecAudioEncoder", "4382": "data2vec/modeling_data2vec_audio.py:Data2VecAudioAdapterLayer", "4383": "data2vec/modeling_data2vec_audio.py:Data2VecAudioAdapter", "4384": "data2vec/modeling_data2vec_audio.py:Data2VecAudioPreTrainedModel", "4385": "data2vec/modeling_data2vec_audio.py:_compute_mask_indices", "4386": "data2vec/modeling_data2vec_audio.py:Data2VecAudioModel", "4387": "data2vec/modeling_data2vec_audio.py:Data2VecAudioForCTC", "4388": "data2vec/modeling_data2vec_audio.py:Data2VecAudioForSequenceClassification", "4389": "data2vec/modeling_data2vec_audio.py:Data2VecAudioForAudioFrameClassification", "4390": "data2vec/modeling_data2vec_audio.py:AMSoftmaxLoss", "4391": "data2vec/modeling_data2vec_audio.py:TDNNLayer", "4392": "data2vec/modeling_data2vec_audio.py:Data2VecAudioForXVector", "4393": "data2vec/modeling_data2vec_text.py:Data2VecTextEmbeddings", "4394": "data2vec/modeling_data2vec_text.py:eager_attention_forward", "4395": "data2vec/modeling_data2vec_text.py:Data2VecTextSelfAttention", "4396": "data2vec/modeling_data2vec_text.py:Data2VecTextCrossAttention", "4397": "data2vec/modeling_data2vec_text.py:Data2VecTextSelfOutput", "4398": "data2vec/modeling_data2vec_text.py:Data2VecTextAttention", "4399": "data2vec/modeling_data2vec_text.py:Data2VecTextIntermediate", "4400": "data2vec/modeling_data2vec_text.py:Data2VecTextOutput", "4401": "data2vec/modeling_data2vec_text.py:Data2VecTextLayer", "4402": "data2vec/modeling_data2vec_text.py:Data2VecTextPreTrainedModel", "4403": "data2vec/modeling_data2vec_text.py:Data2VecTextEncoder", "4404": "data2vec/modeling_data2vec_text.py:Data2VecTextPooler", "4405": "data2vec/modeling_data2vec_text.py:Data2VecTextModel", "4406": "data2vec/modeling_data2vec_text.py:Data2VecTextLMHead", "4407": "data2vec/modeling_data2vec_text.py:Data2VecTextClassificationHead", "4408": "data2vec/modeling_data2vec_text.py:Data2VecTextForCausalLM", "4409": "data2vec/modeling_data2vec_text.py:Data2VecTextForMaskedLM", "4410": "data2vec/modeling_data2vec_text.py:Data2VecTextForSequenceClassification", "4411": "data2vec/modeling_data2vec_text.py:Data2VecTextForMultipleChoice", "4412": "data2vec/modeling_data2vec_text.py:Data2VecTextForTokenClassification", "4413": "data2vec/modeling_data2vec_text.py:Data2VecTextForQuestionAnswering", "4414": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingLayer", "4415": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingPreActResidualLayer", "4416": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingFeatureFusionLayer", "4417": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingFeatureFusionStage", "4418": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingDepthEstimationHead", "4419": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingPreTrainedModel", "4420": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingReassembleLayer", "4421": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingReassembleStage", "4422": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingNeck", "4423": "prompt_depth_anything/modeling_prompt_depth_anything.py:PromptDepthAnythingForDepthEstimation", "4424": "modernbert/modeling_modernbert.py:ApplyRotaryEmbUnpad", "4425": "modernbert/modeling_modernbert.py:apply_rotary_unpadded", "4426": "modernbert/modeling_modernbert.py:ModernBertUnpaddedRotaryEmbedding", "4427": "modernbert/modeling_modernbert.py:ModernBertEmbeddings", "4428": "modernbert/modeling_modernbert.py:ModernBertMLP", "4429": "modernbert/modeling_modernbert.py:ModernBertRotaryEmbedding", "4430": "modernbert/modeling_modernbert.py:rotate_half", "4431": "modernbert/modeling_modernbert.py:apply_rotary_pos_emb", "4432": "modernbert/modeling_modernbert.py:eager_attention_forward", "4433": "modernbert/modeling_modernbert.py:flash_attention_forward", "4434": "modernbert/modeling_modernbert.py:sdpa_attention_forward", "4435": "modernbert/modeling_modernbert.py:ModernBertAttention", "4436": "modernbert/modeling_modernbert.py:ModernBertEncoderLayer", "4437": "modernbert/modeling_modernbert.py:ModernBertPreTrainedModel", "4438": "modernbert/modeling_modernbert.py:_unpad_modernbert_input", "4439": "modernbert/modeling_modernbert.py:_pad_modernbert_output", "4440": "modernbert/modeling_modernbert.py:ModernBertModel", "4441": "modernbert/modeling_modernbert.py:ModernBertPredictionHead", "4442": "modernbert/modeling_modernbert.py:ModernBertForMaskedLM", "4443": "modernbert/modeling_modernbert.py:ModernBertForSequenceClassification", "4444": "modernbert/modeling_modernbert.py:ModernBertForTokenClassification", "4445": "modernbert/modeling_modernbert.py:ModernBertForQuestionAnswering", "4446": "modernbert/modeling_modernbert.py:ModernBertForMultipleChoice", "4447": "ministral/modeling_ministral.py:MinistralMLP", "4448": "ministral/modeling_ministral.py:rotate_half", "4449": "ministral/modeling_ministral.py:apply_rotary_pos_emb", "4450": "ministral/modeling_ministral.py:repeat_kv", "4451": "ministral/modeling_ministral.py:eager_attention_forward", "4452": "ministral/modeling_ministral.py:MinistralAttention", "4453": "ministral/modeling_ministral.py:MinistralRMSNorm", "4454": "ministral/modeling_ministral.py:MinistralDecoderLayer", "4455": "ministral/modeling_ministral.py:MinistralPreTrainedModel", "4456": "ministral/modeling_ministral.py:MinistralRotaryEmbedding", "4457": "ministral/modeling_ministral.py:MinistralModel", "4458": "ministral/modeling_ministral.py:MinistralForCausalLM", "4459": "ministral/modeling_ministral.py:MinistralForSequenceClassification", "4460": "ministral/modeling_ministral.py:MinistralForTokenClassification", "4461": "ministral/modeling_ministral.py:MinistralForQuestionAnswering", "4462": "bark/modeling_bark.py:BarkSelfAttention", "4463": "bark/modeling_bark.py:BarkSelfFlashAttention2", "4464": "bark/modeling_bark.py:BarkMLP", "4465": "bark/modeling_bark.py:BarkBlock", "4466": "bark/modeling_bark.py:BarkPreTrainedModel", "4467": "bark/modeling_bark.py:BarkCausalModel", "4468": "bark/modeling_bark.py:BarkSemanticModel", "4469": "bark/modeling_bark.py:BarkCoarseModel", "4470": "bark/modeling_bark.py:BarkFineModel", "4471": "bark/modeling_bark.py:BarkModel", "4472": "falcon/modeling_falcon.py:FalconLinear", "4473": "falcon/modeling_falcon.py:rotate_half", "4474": "falcon/modeling_falcon.py:apply_rotary_pos_emb", "4475": "falcon/modeling_falcon.py:FalconRotaryEmbedding", "4476": "falcon/modeling_falcon.py:build_alibi_tensor", "4477": "falcon/modeling_falcon.py:dropout_add", "4478": "falcon/modeling_falcon.py:FalconAttention", "4479": "falcon/modeling_falcon.py:FalconFlashAttention2", "4480": "falcon/modeling_falcon.py:FalconMLP", "4481": "falcon/modeling_falcon.py:FalconDecoderLayer", "4482": "falcon/modeling_falcon.py:FalconPreTrainedModel", "4483": "falcon/modeling_falcon.py:FalconModel", "4484": "falcon/modeling_falcon.py:FalconForCausalLM", "4485": "falcon/modeling_falcon.py:FalconForSequenceClassification", "4486": "falcon/modeling_falcon.py:FalconForTokenClassification", "4487": "falcon/modeling_falcon.py:FalconForQuestionAnswering", "4488": "lfm2/modeling_lfm2.py:Lfm2RMSNorm", "4489": "lfm2/modeling_lfm2.py:Lfm2RotaryEmbedding", "4490": "lfm2/modeling_lfm2.py:Lfm2MLP", "4491": "lfm2/modeling_lfm2.py:Lfm2HybridConvCache", "4492": "lfm2/modeling_lfm2.py:rotate_half", "4493": "lfm2/modeling_lfm2.py:apply_rotary_pos_emb", "4494": "lfm2/modeling_lfm2.py:repeat_kv", "4495": "lfm2/modeling_lfm2.py:eager_attention_forward", "4496": "lfm2/modeling_lfm2.py:Lfm2Attention", "4497": "lfm2/modeling_lfm2.py:apply_mask_to_padding_states", "4498": "lfm2/modeling_lfm2.py:Lfm2ShortConv", "4499": "lfm2/modeling_lfm2.py:Lfm2DecoderLayer", "4500": "lfm2/modeling_lfm2.py:Lfm2PreTrainedModel", "4501": "lfm2/modeling_lfm2.py:Lfm2Model", "4502": "lfm2/modeling_lfm2.py:Lfm2ForCausalLM", "4503": "opt/modeling_opt.py:OPTLearnedPositionalEmbedding", "4504": "opt/modeling_opt.py:eager_attention_forward", "4505": "opt/modeling_opt.py:OPTAttention", "4506": "opt/modeling_opt.py:OPTDecoderLayer", "4507": "opt/modeling_opt.py:OPTPreTrainedModel", "4508": "opt/modeling_opt.py:OPTDecoder", "4509": "opt/modeling_opt.py:OPTModel", "4510": "opt/modeling_opt.py:OPTForCausalLM", "4511": "opt/modeling_opt.py:OPTForSequenceClassification", "4512": "opt/modeling_opt.py:OPTForQuestionAnswering", "4513": "m2m_100/modeling_m2m_100.py:shift_tokens_right", "4514": "m2m_100/modeling_m2m_100.py:M2M100ScaledWordEmbedding", "4515": "m2m_100/modeling_m2m_100.py:M2M100SinusoidalPositionalEmbedding", "4516": "m2m_100/modeling_m2m_100.py:eager_attention_forward", "4517": "m2m_100/modeling_m2m_100.py:M2M100Attention", "4518": "m2m_100/modeling_m2m_100.py:M2M100EncoderLayer", "4519": "m2m_100/modeling_m2m_100.py:M2M100DecoderLayer", "4520": "m2m_100/modeling_m2m_100.py:M2M100PreTrainedModel", "4521": "m2m_100/modeling_m2m_100.py:M2M100Encoder", "4522": "m2m_100/modeling_m2m_100.py:M2M100Decoder", "4523": "m2m_100/modeling_m2m_100.py:M2M100Model", "4524": "m2m_100/modeling_m2m_100.py:M2M100ForConditionalGeneration", "4525": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboEncoderOutput", "4526": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboDecoderOutput", "4527": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboObjectDetectionOutput", "4528": "omdet_turbo/modeling_omdet_turbo.py:MultiScaleDeformableAttention", "4529": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboLRUCache", "4530": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboLanguageBackbone", "4531": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboVisionBackbone", "4532": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboMultiscaleDeformableAttention", "4533": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboConvNormLayer", "4534": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboRepVggBlock", "4535": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboCSPRepLayer", "4536": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboMultiheadAttention", "4537": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboEncoderLayer", "4538": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboEncoder", "4539": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboHybridEncoder", "4540": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboMLPWithDropout", "4541": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboMLP", "4542": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboResidualLayer", "4543": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboTaskEncoder", "4544": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboDeformableTransformerDecoderLayer", "4545": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboPreTrainedModel", "4546": "omdet_turbo/modeling_omdet_turbo.py:_cosine_similarity_scaled", "4547": "omdet_turbo/modeling_omdet_turbo.py:get_class_similarity", "4548": "omdet_turbo/modeling_omdet_turbo.py:_inverse_sigmoid", "4549": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboDecoder", "4550": "omdet_turbo/modeling_omdet_turbo.py:OmDetTurboForObjectDetection", "4551": "blip/modeling_blip.py:contrastive_loss", "4552": "blip/modeling_blip.py:blip_loss", "4553": "blip/modeling_blip.py:BlipForConditionalGenerationModelOutput", "4554": "blip/modeling_blip.py:BlipTextVisionModelOutput", "4555": "blip/modeling_blip.py:BlipImageTextMatchingModelOutput", "4556": "blip/modeling_blip.py:BlipOutput", "4557": "blip/modeling_blip.py:BlipVisionEmbeddings", "4558": "blip/modeling_blip.py:BlipTextEmbeddings", "4559": "blip/modeling_blip.py:BlipAttention", "4560": "blip/modeling_blip.py:BlipMLP", "4561": "blip/modeling_blip.py:BlipEncoderLayer", "4562": "blip/modeling_blip.py:BlipPreTrainedModel", "4563": "blip/modeling_blip.py:BlipEncoder", "4564": "blip/modeling_blip.py:BlipVisionModel", "4565": "blip/modeling_blip.py:BlipModel", "4566": "blip/modeling_blip.py:BlipForConditionalGeneration", "4567": "blip/modeling_blip.py:BlipForQuestionAnswering", "4568": "blip/modeling_blip.py:BlipForImageTextRetrieval", "4569": "blip/modeling_blip_text.py:BlipTextEmbeddings", "4570": "blip/modeling_blip_text.py:BlipTextSelfAttention", "4571": "blip/modeling_blip_text.py:BlipTextSelfOutput", "4572": "blip/modeling_blip_text.py:BlipTextAttention", "4573": "blip/modeling_blip_text.py:BlipTextIntermediate", "4574": "blip/modeling_blip_text.py:BlipTextOutput", "4575": "blip/modeling_blip_text.py:BlipTextLayer", "4576": "blip/modeling_blip_text.py:BlipTextEncoder", "4577": "blip/modeling_blip_text.py:BlipTextPooler", "4578": "blip/modeling_blip_text.py:BlipTextPredictionHeadTransform", "4579": "blip/modeling_blip_text.py:BlipTextLMPredictionHead", "4580": "blip/modeling_blip_text.py:BlipTextOnlyMLMHead", "4581": "blip/modeling_blip_text.py:BlipTextPreTrainedModel", "4582": "blip/modeling_blip_text.py:BlipTextModel", "4583": "blip/modeling_blip_text.py:BlipTextLMHeadModel", "4584": "sew/modeling_sew.py:SEWNoLayerNormConvLayer", "4585": "sew/modeling_sew.py:SEWLayerNormConvLayer", "4586": "sew/modeling_sew.py:SEWGroupNormConvLayer", "4587": "sew/modeling_sew.py:SEWPositionalConvEmbedding", "4588": "sew/modeling_sew.py:SEWSamePadLayer", "4589": "sew/modeling_sew.py:SEWUpsampling", "4590": "sew/modeling_sew.py:SEWFeatureEncoder", "4591": "sew/modeling_sew.py:eager_attention_forward", "4592": "sew/modeling_sew.py:SEWAttention", "4593": "sew/modeling_sew.py:SEWFeedForward", "4594": "sew/modeling_sew.py:SEWEncoderLayer", "4595": "sew/modeling_sew.py:SEWEncoder", "4596": "sew/modeling_sew.py:SEWPreTrainedModel", "4597": "sew/modeling_sew.py:_compute_mask_indices", "4598": "sew/modeling_sew.py:SEWModel", "4599": "sew/modeling_sew.py:SEWForCTC", "4600": "sew/modeling_sew.py:SEWForSequenceClassification", "4601": "gpt_oss/modeling_gpt_oss.py:GptOssRMSNorm", "4602": "gpt_oss/modeling_gpt_oss.py:GptOssExperts", "4603": "gpt_oss/modeling_gpt_oss.py:GptOssTopKRouter", "4604": "gpt_oss/modeling_gpt_oss.py:GptOssMLP", "4605": "gpt_oss/modeling_gpt_oss.py:GptOssRotaryEmbedding", "4606": "gpt_oss/modeling_gpt_oss.py:repeat_kv", "4607": "gpt_oss/modeling_gpt_oss.py:_apply_rotary_emb", "4608": "gpt_oss/modeling_gpt_oss.py:apply_rotary_pos_emb", "4609": "gpt_oss/modeling_gpt_oss.py:eager_attention_forward", "4610": "gpt_oss/modeling_gpt_oss.py:GptOssAttention", "4611": "gpt_oss/modeling_gpt_oss.py:GptOssDecoderLayer", "4612": "gpt_oss/modeling_gpt_oss.py:GptOssPreTrainedModel", "4613": "gpt_oss/modeling_gpt_oss.py:GptOssModel", "4614": "gpt_oss/modeling_gpt_oss.py:load_balancing_loss_func", "4615": "gpt_oss/modeling_gpt_oss.py:GptOssForCausalLM", "4616": "gpt_oss/modeling_gpt_oss.py:GptOssForSequenceClassification", "4617": "gpt_oss/modeling_gpt_oss.py:GptOssForTokenClassification", "4618": "hubert/modeling_hubert.py:HubertPositionalConvEmbedding", "4619": "hubert/modeling_hubert.py:HubertSamePadLayer", "4620": "hubert/modeling_hubert.py:HubertNoLayerNormConvLayer", "4621": "hubert/modeling_hubert.py:HubertLayerNormConvLayer", "4622": "hubert/modeling_hubert.py:HubertGroupNormConvLayer", "4623": "hubert/modeling_hubert.py:HubertFeatureEncoder", "4624": "hubert/modeling_hubert.py:HubertFeatureProjection", "4625": "hubert/modeling_hubert.py:eager_attention_forward", "4626": "hubert/modeling_hubert.py:HubertAttention", "4627": "hubert/modeling_hubert.py:HubertFeedForward", "4628": "hubert/modeling_hubert.py:HubertEncoderLayer", "4629": "hubert/modeling_hubert.py:HubertEncoder", "4630": "hubert/modeling_hubert.py:HubertAttnAdapterLayer", "4631": "hubert/modeling_hubert.py:HubertEncoderLayerStableLayerNorm", "4632": "hubert/modeling_hubert.py:HubertEncoderStableLayerNorm", "4633": "hubert/modeling_hubert.py:HubertPreTrainedModel", "4634": "hubert/modeling_hubert.py:_compute_mask_indices", "4635": "hubert/modeling_hubert.py:HubertModel", "4636": "hubert/modeling_hubert.py:HubertForCTC", "4637": "hubert/modeling_hubert.py:HubertForSequenceClassification", "4638": "swin/modeling_swin.py:SwinEncoderOutput", "4639": "swin/modeling_swin.py:SwinModelOutput", "4640": "swin/modeling_swin.py:SwinMaskedImageModelingOutput", "4641": "swin/modeling_swin.py:SwinImageClassifierOutput", "4642": "swin/modeling_swin.py:window_partition", "4643": "swin/modeling_swin.py:window_reverse", "4644": "swin/modeling_swin.py:SwinEmbeddings", "4645": "swin/modeling_swin.py:SwinPatchEmbeddings", "4646": "swin/modeling_swin.py:SwinPatchMerging", "4647": "swin/modeling_swin.py:drop_path", "4648": "swin/modeling_swin.py:SwinDropPath", "4649": "swin/modeling_swin.py:SwinSelfAttention", "4650": "swin/modeling_swin.py:SwinSelfOutput", "4651": "swin/modeling_swin.py:SwinAttention", "4652": "swin/modeling_swin.py:SwinIntermediate", "4653": "swin/modeling_swin.py:SwinOutput", "4654": "swin/modeling_swin.py:SwinLayer", "4655": "swin/modeling_swin.py:SwinStage", "4656": "swin/modeling_swin.py:SwinEncoder", "4657": "swin/modeling_swin.py:SwinPreTrainedModel", "4658": "swin/modeling_swin.py:SwinModel", "4659": "swin/modeling_swin.py:SwinForMaskedImageModeling", "4660": "swin/modeling_swin.py:SwinForImageClassification", "4661": "swin/modeling_swin.py:SwinBackbone", "4662": "squeezebert/modeling_squeezebert.py:SqueezeBertEmbeddings", "4663": "squeezebert/modeling_squeezebert.py:MatMulWrapper", "4664": "squeezebert/modeling_squeezebert.py:SqueezeBertLayerNorm", "4665": "squeezebert/modeling_squeezebert.py:ConvDropoutLayerNorm", "4666": "squeezebert/modeling_squeezebert.py:ConvActivation", "4667": "squeezebert/modeling_squeezebert.py:SqueezeBertSelfAttention", "4668": "squeezebert/modeling_squeezebert.py:SqueezeBertModule", "4669": "squeezebert/modeling_squeezebert.py:SqueezeBertEncoder", "4670": "squeezebert/modeling_squeezebert.py:SqueezeBertPooler", "4671": "squeezebert/modeling_squeezebert.py:SqueezeBertPredictionHeadTransform", "4672": "squeezebert/modeling_squeezebert.py:SqueezeBertLMPredictionHead", "4673": "squeezebert/modeling_squeezebert.py:SqueezeBertOnlyMLMHead", "4674": "squeezebert/modeling_squeezebert.py:SqueezeBertPreTrainedModel", "4675": "squeezebert/modeling_squeezebert.py:SqueezeBertModel", "4676": "squeezebert/modeling_squeezebert.py:SqueezeBertForMaskedLM", "4677": "squeezebert/modeling_squeezebert.py:SqueezeBertForSequenceClassification", "4678": "squeezebert/modeling_squeezebert.py:SqueezeBertForMultipleChoice", "4679": "squeezebert/modeling_squeezebert.py:SqueezeBertForTokenClassification", "4680": "squeezebert/modeling_squeezebert.py:SqueezeBertForQuestionAnswering", "4681": "lfm2_vl/modeling_lfm2_vl.py:Lfm2VlMultiModalProjector", "4682": "lfm2_vl/modeling_lfm2_vl.py:Lfm2VlPreTrainedModel", "4683": "lfm2_vl/modeling_lfm2_vl.py:Lfm2VlCausalLMOutputWithPast", "4684": "lfm2_vl/modeling_lfm2_vl.py:Lfm2VlModelOutputWithPast", "4685": "lfm2_vl/modeling_lfm2_vl.py:Lfm2VlModel", "4686": "lfm2_vl/modeling_lfm2_vl.py:Lfm2VlForConditionalGeneration", "4687": "superpoint/modeling_superpoint.py:remove_keypoints_from_borders", "4688": "superpoint/modeling_superpoint.py:top_k_keypoints", "4689": "superpoint/modeling_superpoint.py:simple_nms", "4690": "superpoint/modeling_superpoint.py:SuperPointKeypointDescriptionOutput", "4691": "superpoint/modeling_superpoint.py:SuperPointConvBlock", "4692": "superpoint/modeling_superpoint.py:SuperPointEncoder", "4693": "superpoint/modeling_superpoint.py:SuperPointInterestPointDecoder", "4694": "superpoint/modeling_superpoint.py:SuperPointDescriptorDecoder", "4695": "superpoint/modeling_superpoint.py:SuperPointPreTrainedModel", "4696": "superpoint/modeling_superpoint.py:SuperPointForKeypointDetection", "4697": "gemma2/modeling_gemma2.py:Gemma2RMSNorm", "4698": "gemma2/modeling_gemma2.py:Gemma2MLP", "4699": "gemma2/modeling_gemma2.py:Gemma2RotaryEmbedding", "4700": "gemma2/modeling_gemma2.py:rotate_half", "4701": "gemma2/modeling_gemma2.py:apply_rotary_pos_emb", "4702": "gemma2/modeling_gemma2.py:repeat_kv", "4703": "gemma2/modeling_gemma2.py:eager_attention_forward", "4704": "gemma2/modeling_gemma2.py:Gemma2Attention", "4705": "gemma2/modeling_gemma2.py:Gemma2DecoderLayer", "4706": "gemma2/modeling_gemma2.py:Gemma2PreTrainedModel", "4707": "gemma2/modeling_gemma2.py:Gemma2Model", "4708": "gemma2/modeling_gemma2.py:Gemma2ForCausalLM", "4709": "gemma2/modeling_gemma2.py:Gemma2ForSequenceClassification", "4710": "gemma2/modeling_gemma2.py:Gemma2ForTokenClassification", "4711": "git/modeling_git.py:GitVisionModelOutput", "4712": "git/modeling_git.py:GitEmbeddings", "4713": "git/modeling_git.py:GitSelfAttention", "4714": "git/modeling_git.py:GitSelfOutput", "4715": "git/modeling_git.py:GitAttention", "4716": "git/modeling_git.py:GitIntermediate", "4717": "git/modeling_git.py:GitOutput", "4718": "git/modeling_git.py:GitLayer", "4719": "git/modeling_git.py:GitEncoder", "4720": "git/modeling_git.py:GitPreTrainedModel", "4721": "git/modeling_git.py:GitVisionEmbeddings", "4722": "git/modeling_git.py:GitVisionMLP", "4723": "git/modeling_git.py:eager_attention_forward", "4724": "git/modeling_git.py:GitVisionAttention", "4725": "git/modeling_git.py:GitVisionEncoderLayer", "4726": "git/modeling_git.py:GitVisionEncoder", "4727": "git/modeling_git.py:GitVisionTransformer", "4728": "git/modeling_git.py:GitVisionModel", "4729": "git/modeling_git.py:GitProjection", "4730": "git/modeling_git.py:GitModel", "4731": "git/modeling_git.py:GitForCausalLM", "4732": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetConvLayer", "4733": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetEmbeddings", "4734": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetShortCut", "4735": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetBasicLayer", "4736": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetBottleNeckLayer", "4737": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetStage", "4738": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetEncoder", "4739": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetPreTrainedModel", "4740": "rt_detr/modeling_rt_detr_resnet.py:RTDetrResNetBackbone", "4741": "rt_detr/modeling_rt_detr.py:MultiScaleDeformableAttention", "4742": "rt_detr/modeling_rt_detr.py:RTDetrDecoderOutput", "4743": "rt_detr/modeling_rt_detr.py:RTDetrModelOutput", "4744": "rt_detr/modeling_rt_detr.py:RTDetrObjectDetectionOutput", "4745": "rt_detr/modeling_rt_detr.py:_get_clones", "4746": "rt_detr/modeling_rt_detr.py:inverse_sigmoid", "4747": "rt_detr/modeling_rt_detr.py:RTDetrFrozenBatchNorm2d", "4748": "rt_detr/modeling_rt_detr.py:replace_batch_norm", "4749": "rt_detr/modeling_rt_detr.py:get_contrastive_denoising_training_group", "4750": "rt_detr/modeling_rt_detr.py:RTDetrConvEncoder", "4751": "rt_detr/modeling_rt_detr.py:RTDetrConvNormLayer", "4752": "rt_detr/modeling_rt_detr.py:RTDetrEncoderLayer", "4753": "rt_detr/modeling_rt_detr.py:RTDetrRepVggBlock", "4754": "rt_detr/modeling_rt_detr.py:RTDetrCSPRepLayer", "4755": "rt_detr/modeling_rt_detr.py:RTDetrMultiscaleDeformableAttention", "4756": "rt_detr/modeling_rt_detr.py:RTDetrMultiheadAttention", "4757": "rt_detr/modeling_rt_detr.py:RTDetrDecoderLayer", "4758": "rt_detr/modeling_rt_detr.py:RTDetrPreTrainedModel", "4759": "rt_detr/modeling_rt_detr.py:RTDetrEncoder", "4760": "rt_detr/modeling_rt_detr.py:RTDetrHybridEncoder", "4761": "rt_detr/modeling_rt_detr.py:RTDetrDecoder", "4762": "rt_detr/modeling_rt_detr.py:RTDetrMLPPredictionHead", "4763": "rt_detr/modeling_rt_detr.py:RTDetrModel", "4764": "rt_detr/modeling_rt_detr.py:RTDetrForObjectDetection", "4765": "idefics3/modeling_idefics3.py:Idefics3BaseModelOutputWithPast", "4766": "idefics3/modeling_idefics3.py:Idefics3CausalLMOutputWithPast", "4767": "idefics3/modeling_idefics3.py:Idefics3VisionEmbeddings", "4768": "idefics3/modeling_idefics3.py:eager_attention_forward", "4769": "idefics3/modeling_idefics3.py:Idefics3VisionAttention", "4770": "idefics3/modeling_idefics3.py:Idefics3VisionMLP", "4771": "idefics3/modeling_idefics3.py:Idefics3SimpleMLP", "4772": "idefics3/modeling_idefics3.py:Idefics3EncoderLayer", "4773": "idefics3/modeling_idefics3.py:Idefics3Encoder", "4774": "idefics3/modeling_idefics3.py:repeat_kv", "4775": "idefics3/modeling_idefics3.py:Idefics3RMSNorm", "4776": "idefics3/modeling_idefics3.py:Idefics3Connector", "4777": "idefics3/modeling_idefics3.py:Idefics3PreTrainedModel", "4778": "idefics3/modeling_idefics3.py:Idefics3VisionTransformer", "4779": "idefics3/modeling_idefics3.py:Idefics3Model", "4780": "idefics3/modeling_idefics3.py:Idefics3ForConditionalGeneration", "4781": "idefics2/modeling_idefics2.py:Idefics2BaseModelOutputWithPast", "4782": "idefics2/modeling_idefics2.py:Idefics2CausalLMOutputWithPast", "4783": "idefics2/modeling_idefics2.py:Idefics2VisionEmbeddings", "4784": "idefics2/modeling_idefics2.py:eager_attention_forward", "4785": "idefics2/modeling_idefics2.py:Idefics2VisionAttention", "4786": "idefics2/modeling_idefics2.py:Idefics2VisionMLP", "4787": "idefics2/modeling_idefics2.py:Idefics2MLP", "4788": "idefics2/modeling_idefics2.py:Idefics2MultiheadAttentionPoolingHead", "4789": "idefics2/modeling_idefics2.py:Idefics2EncoderLayer", "4790": "idefics2/modeling_idefics2.py:Idefics2Encoder", "4791": "idefics2/modeling_idefics2.py:Idefics2PreTrainedModel", "4792": "idefics2/modeling_idefics2.py:Idefics2VisionTransformer", "4793": "idefics2/modeling_idefics2.py:repeat_kv", "4794": "idefics2/modeling_idefics2.py:Idefics2RMSNorm", "4795": "idefics2/modeling_idefics2.py:Idefics2PerceiverAttention", "4796": "idefics2/modeling_idefics2.py:Idefics2PerceiverLayer", "4797": "idefics2/modeling_idefics2.py:Idefics2PerceiverResampler", "4798": "idefics2/modeling_idefics2.py:Idefics2Connector", "4799": "idefics2/modeling_idefics2.py:Idefics2Model", "4800": "idefics2/modeling_idefics2.py:Idefics2ForConditionalGeneration", "4801": "d_fine/modeling_d_fine.py:multi_scale_deformable_attention_v2", "4802": "d_fine/modeling_d_fine.py:DFineMultiscaleDeformableAttention", "4803": "d_fine/modeling_d_fine.py:DFineGate", "4804": "d_fine/modeling_d_fine.py:DFineMultiheadAttention", "4805": "d_fine/modeling_d_fine.py:DFineDecoderLayer", "4806": "d_fine/modeling_d_fine.py:DFinePreTrainedModel", "4807": "d_fine/modeling_d_fine.py:DFineIntegral", "4808": "d_fine/modeling_d_fine.py:DFineDecoderOutput", "4809": "d_fine/modeling_d_fine.py:inverse_sigmoid", "4810": "d_fine/modeling_d_fine.py:weighting_function", "4811": "d_fine/modeling_d_fine.py:distance2bbox", "4812": "d_fine/modeling_d_fine.py:DFineDecoder", "4813": "d_fine/modeling_d_fine.py:DFineModelOutput", "4814": "d_fine/modeling_d_fine.py:DFineFrozenBatchNorm2d", "4815": "d_fine/modeling_d_fine.py:replace_batch_norm", "4816": "d_fine/modeling_d_fine.py:DFineConvEncoder", "4817": "d_fine/modeling_d_fine.py:get_contrastive_denoising_training_group", "4818": "d_fine/modeling_d_fine.py:DFineModel", "4819": "d_fine/modeling_d_fine.py:DFineObjectDetectionOutput", "4820": "d_fine/modeling_d_fine.py:DFineForObjectDetection", "4821": "d_fine/modeling_d_fine.py:DFineMLPPredictionHead", "4822": "d_fine/modeling_d_fine.py:DFineMLP", "4823": "d_fine/modeling_d_fine.py:DFineLQE", "4824": "d_fine/modeling_d_fine.py:DFineConvNormLayer", "4825": "d_fine/modeling_d_fine.py:DFineRepVggBlock", "4826": "d_fine/modeling_d_fine.py:DFineCSPRepLayer", "4827": "d_fine/modeling_d_fine.py:DFineRepNCSPELAN4", "4828": "d_fine/modeling_d_fine.py:DFineSCDown", "4829": "d_fine/modeling_d_fine.py:DFineEncoderLayer", "4830": "d_fine/modeling_d_fine.py:DFineEncoder", "4831": "d_fine/modeling_d_fine.py:DFineHybridEncoder", "4832": "mistral3/modeling_mistral3.py:Mistral3RMSNorm", "4833": "mistral3/modeling_mistral3.py:Mistral3PatchMerger", "4834": "mistral3/modeling_mistral3.py:Mistral3MultiModalProjector", "4835": "mistral3/modeling_mistral3.py:Mistral3CausalLMOutputWithPast", "4836": "mistral3/modeling_mistral3.py:Mistral3ModelOutputWithPast", "4837": "mistral3/modeling_mistral3.py:Mistral3PreTrainedModel", "4838": "mistral3/modeling_mistral3.py:Mistral3Model", "4839": "mistral3/modeling_mistral3.py:Mistral3ForConditionalGeneration", "4840": "imagegpt/modeling_imagegpt.py:ImageGPTLayerNorm", "4841": "imagegpt/modeling_imagegpt.py:ImageGPTAttention", "4842": "imagegpt/modeling_imagegpt.py:ImageGPTMLP", "4843": "imagegpt/modeling_imagegpt.py:ImageGPTBlock", "4844": "imagegpt/modeling_imagegpt.py:ImageGPTPreTrainedModel", "4845": "imagegpt/modeling_imagegpt.py:ImageGPTModel", "4846": "imagegpt/modeling_imagegpt.py:ImageGPTForCausalImageModeling", "4847": "imagegpt/modeling_imagegpt.py:ImageGPTForImageClassification", "4848": "moshi/modeling_moshi.py:MoshiConditionalGenerationGenerateOutput", "4849": "moshi/modeling_moshi.py:MoshiCausalLMOutputWithPast", "4850": "moshi/modeling_moshi.py:MoshiConditionalGenerationOutputWithPast", "4851": "moshi/modeling_moshi.py:MoshiUnconditionalInput", "4852": "moshi/modeling_moshi.py:MoshiRMSNorm", "4853": "moshi/modeling_moshi.py:MoshiFlexibleLinear", "4854": "moshi/modeling_moshi.py:MoshiLinear", "4855": "moshi/modeling_moshi.py:MoshiRotaryEmbedding", "4856": "moshi/modeling_moshi.py:rotate_half", "4857": "moshi/modeling_moshi.py:apply_rotary_pos_emb", "4858": "moshi/modeling_moshi.py:MoshiGatingMLP", "4859": "moshi/modeling_moshi.py:repeat_kv", "4860": "moshi/modeling_moshi.py:MoshiAttention", "4861": "moshi/modeling_moshi.py:MoshiFlashAttention2", "4862": "moshi/modeling_moshi.py:MoshiSdpaAttention", "4863": "moshi/modeling_moshi.py:MoshiDecoderLayer", "4864": "moshi/modeling_moshi.py:MoshiPreTrainedModel", "4865": "moshi/modeling_moshi.py:MoshiDepthDecoder", "4866": "moshi/modeling_moshi.py:MoshiModel", "4867": "moshi/modeling_moshi.py:MoshiForCausalLM", "4868": "moshi/modeling_moshi.py:MoshiForConditionalGeneration", "4869": "shieldgemma2/modeling_shieldgemma2.py:ShieldGemma2ImageClassifierOutputWithNoAttention", "4870": "shieldgemma2/modeling_shieldgemma2.py:ShieldGemma2ForImageClassification", "4871": "vision_text_dual_encoder/modeling_vision_text_dual_encoder.py:contrastive_loss", "4872": "vision_text_dual_encoder/modeling_vision_text_dual_encoder.py:clip_loss", "4873": "vision_text_dual_encoder/modeling_vision_text_dual_encoder.py:VisionTextDualEncoderModel", "4874": "distilbert/modeling_distilbert.py:create_sinusoidal_embeddings", "4875": "distilbert/modeling_distilbert.py:_create_sinusoidal_embeddings", "4876": "distilbert/modeling_distilbert.py:Embeddings", "4877": "distilbert/modeling_distilbert.py:MultiHeadSelfAttention", "4878": "distilbert/modeling_distilbert.py:DistilBertFlashAttention2", "4879": "distilbert/modeling_distilbert.py:DistilBertSdpaAttention", "4880": "distilbert/modeling_distilbert.py:FFN", "4881": "distilbert/modeling_distilbert.py:TransformerBlock", "4882": "distilbert/modeling_distilbert.py:Transformer", "4883": "distilbert/modeling_distilbert.py:DistilBertPreTrainedModel", "4884": "distilbert/modeling_distilbert.py:DistilBertModel", "4885": "distilbert/modeling_distilbert.py:DistilBertForMaskedLM", "4886": "distilbert/modeling_distilbert.py:DistilBertForSequenceClassification", "4887": "distilbert/modeling_distilbert.py:DistilBertForQuestionAnswering", "4888": "distilbert/modeling_distilbert.py:DistilBertForTokenClassification", "4889": "distilbert/modeling_distilbert.py:DistilBertForMultipleChoice", "4890": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderEmbeddings", "4891": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderMLP", "4892": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderRotaryEmbedding", "4893": "modernbert_decoder/modeling_modernbert_decoder.py:rotate_half", "4894": "modernbert_decoder/modeling_modernbert_decoder.py:apply_rotary_pos_emb", "4895": "modernbert_decoder/modeling_modernbert_decoder.py:eager_attention_forward", "4896": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderAttention", "4897": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderLayer", "4898": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderPredictionHead", "4899": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderPreTrainedModel", "4900": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderModel", "4901": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderForCausalLM", "4902": "modernbert_decoder/modeling_modernbert_decoder.py:ModernBertDecoderForSequenceClassification", "4903": "deit/modeling_deit.py:DeiTEmbeddings", "4904": "deit/modeling_deit.py:DeiTPatchEmbeddings", "4905": "deit/modeling_deit.py:eager_attention_forward", "4906": "deit/modeling_deit.py:DeiTSelfAttention", "4907": "deit/modeling_deit.py:DeiTSelfOutput", "4908": "deit/modeling_deit.py:DeiTAttention", "4909": "deit/modeling_deit.py:DeiTIntermediate", "4910": "deit/modeling_deit.py:DeiTOutput", "4911": "deit/modeling_deit.py:DeiTLayer", "4912": "deit/modeling_deit.py:DeiTEncoder", "4913": "deit/modeling_deit.py:DeiTPreTrainedModel", "4914": "deit/modeling_deit.py:DeiTModel", "4915": "deit/modeling_deit.py:DeiTPooler", "4916": "deit/modeling_deit.py:DeiTForMaskedImageModeling", "4917": "deit/modeling_deit.py:DeiTForImageClassification", "4918": "deit/modeling_deit.py:DeiTForImageClassificationWithTeacherOutput", "4919": "deit/modeling_deit.py:DeiTForImageClassificationWithTeacher", "4920": "aria/modeling_aria.py:AriaTextRMSNorm", "4921": "aria/modeling_aria.py:AriaProjectorMLP", "4922": "aria/modeling_aria.py:AriaCrossAttention", "4923": "aria/modeling_aria.py:AriaProjector", "4924": "aria/modeling_aria.py:AriaSharedExpertsMLP", "4925": "aria/modeling_aria.py:sequential_experts_gemm", "4926": "aria/modeling_aria.py:AriaGroupedExpertsGemm", "4927": "aria/modeling_aria.py:AriaGroupedExpertsMLP", "4928": "aria/modeling_aria.py:AriaTextMoELayer", "4929": "aria/modeling_aria.py:rotate_half", "4930": "aria/modeling_aria.py:apply_rotary_pos_emb", "4931": "aria/modeling_aria.py:repeat_kv", "4932": "aria/modeling_aria.py:eager_attention_forward", "4933": "aria/modeling_aria.py:AriaTextAttention", "4934": "aria/modeling_aria.py:AriaTextDecoderLayer", "4935": "aria/modeling_aria.py:AriaTextPreTrainedModel", "4936": "aria/modeling_aria.py:AriaPreTrainedModel", "4937": "aria/modeling_aria.py:AriaTextRotaryEmbedding", "4938": "aria/modeling_aria.py:AriaTextModel", "4939": "aria/modeling_aria.py:AriaTextForCausalLM", "4940": "aria/modeling_aria.py:AriaCausalLMOutputWithPast", "4941": "aria/modeling_aria.py:AriaModelOutputWithPast", "4942": "aria/modeling_aria.py:AriaModel", "4943": "aria/modeling_aria.py:AriaForConditionalGeneration", "4944": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1RMSNorm", "4945": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1MLP", "4946": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:rotate_half", "4947": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:apply_rotary_pos_emb", "4948": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:repeat_kv", "4949": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:eager_attention_forward", "4950": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1Attention", "4951": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1DecoderLayer", "4952": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1PreTrainedModel", "4953": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1RotaryEmbedding", "4954": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1Model", "4955": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1ForCausalLM", "4956": "hunyuan_v1_dense/modeling_hunyuan_v1_dense.py:HunYuanDenseV1ForSequenceClassification", "4957": "siglip2/modeling_siglip2.py:Siglip2VisionOutput", "4958": "siglip2/modeling_siglip2.py:Siglip2TextOutput", "4959": "siglip2/modeling_siglip2.py:Siglip2Output", "4960": "siglip2/modeling_siglip2.py:Siglip2VisionEmbeddings", "4961": "siglip2/modeling_siglip2.py:eager_attention_forward", "4962": "siglip2/modeling_siglip2.py:Siglip2Attention", "4963": "siglip2/modeling_siglip2.py:Siglip2MLP", "4964": "siglip2/modeling_siglip2.py:Siglip2EncoderLayer", "4965": "siglip2/modeling_siglip2.py:Siglip2Encoder", "4966": "siglip2/modeling_siglip2.py:Siglip2VisionTransformer", "4967": "siglip2/modeling_siglip2.py:_trunc_normal_", "4968": "siglip2/modeling_siglip2.py:trunc_normal_tf_", "4969": "siglip2/modeling_siglip2.py:variance_scaling_", "4970": "siglip2/modeling_siglip2.py:lecun_normal_", "4971": "siglip2/modeling_siglip2.py:default_flax_embed_init", "4972": "siglip2/modeling_siglip2.py:Siglip2PreTrainedModel", "4973": "siglip2/modeling_siglip2.py:Siglip2TextEmbeddings", "4974": "siglip2/modeling_siglip2.py:Siglip2TextTransformer", "4975": "siglip2/modeling_siglip2.py:Siglip2TextModel", "4976": "siglip2/modeling_siglip2.py:Siglip2MultiheadAttentionPoolingHead", "4977": "siglip2/modeling_siglip2.py:Siglip2VisionModel", "4978": "siglip2/modeling_siglip2.py:Siglip2Model", "4979": "siglip2/modeling_siglip2.py:Siglip2ForImageClassification", "4980": "deberta_v2/modeling_deberta_v2.py:DebertaV2SelfOutput", "4981": "deberta_v2/modeling_deberta_v2.py:make_log_bucket_position", "4982": "deberta_v2/modeling_deberta_v2.py:build_relative_position", "4983": "deberta_v2/modeling_deberta_v2.py:c2p_dynamic_expand", "4984": "deberta_v2/modeling_deberta_v2.py:p2c_dynamic_expand", "4985": "deberta_v2/modeling_deberta_v2.py:pos_dynamic_expand", "4986": "deberta_v2/modeling_deberta_v2.py:scaled_size_sqrt", "4987": "deberta_v2/modeling_deberta_v2.py:build_rpos", "4988": "deberta_v2/modeling_deberta_v2.py:DisentangledSelfAttention", "4989": "deberta_v2/modeling_deberta_v2.py:DebertaV2Attention", "4990": "deberta_v2/modeling_deberta_v2.py:DebertaV2Intermediate", "4991": "deberta_v2/modeling_deberta_v2.py:DebertaV2Output", "4992": "deberta_v2/modeling_deberta_v2.py:DebertaV2Layer", "4993": "deberta_v2/modeling_deberta_v2.py:ConvLayer", "4994": "deberta_v2/modeling_deberta_v2.py:DebertaV2Embeddings", "4995": "deberta_v2/modeling_deberta_v2.py:DebertaV2Encoder", "4996": "deberta_v2/modeling_deberta_v2.py:DebertaV2PreTrainedModel", "4997": "deberta_v2/modeling_deberta_v2.py:DebertaV2Model", "4998": "deberta_v2/modeling_deberta_v2.py:LegacyDebertaV2PredictionHeadTransform", "4999": "deberta_v2/modeling_deberta_v2.py:LegacyDebertaV2LMPredictionHead", "5000": "deberta_v2/modeling_deberta_v2.py:LegacyDebertaV2OnlyMLMHead", "5001": "deberta_v2/modeling_deberta_v2.py:DebertaV2LMPredictionHead", "5002": "deberta_v2/modeling_deberta_v2.py:DebertaV2OnlyMLMHead", "5003": "deberta_v2/modeling_deberta_v2.py:DebertaV2ForMaskedLM", "5004": "deberta_v2/modeling_deberta_v2.py:ContextPooler", "5005": "deberta_v2/modeling_deberta_v2.py:DebertaV2ForSequenceClassification", "5006": "deberta_v2/modeling_deberta_v2.py:DebertaV2ForTokenClassification", "5007": "deberta_v2/modeling_deberta_v2.py:DebertaV2ForQuestionAnswering", "5008": "deberta_v2/modeling_deberta_v2.py:DebertaV2ForMultipleChoice", "5009": "auto/modeling_auto.py:AutoModelForMaskGeneration", "5010": "auto/modeling_auto.py:AutoModelForKeypointDetection", "5011": "auto/modeling_auto.py:AutoModelForKeypointMatching", "5012": "auto/modeling_auto.py:AutoModelForTextEncoding", "5013": "auto/modeling_auto.py:AutoModelForImageToImage", "5014": "auto/modeling_auto.py:AutoModel", "5015": "auto/modeling_auto.py:AutoModelForPreTraining", "5016": "auto/modeling_auto.py:_AutoModelWithLMHead", "5017": "auto/modeling_auto.py:AutoModelForCausalLM", "5018": "auto/modeling_auto.py:AutoModelForMaskedLM", "5019": "auto/modeling_auto.py:AutoModelForSeq2SeqLM", "5020": "auto/modeling_auto.py:AutoModelForSequenceClassification", "5021": "auto/modeling_auto.py:AutoModelForQuestionAnswering", "5022": "auto/modeling_auto.py:AutoModelForTableQuestionAnswering", "5023": "auto/modeling_auto.py:AutoModelForVisualQuestionAnswering", "5024": "auto/modeling_auto.py:AutoModelForDocumentQuestionAnswering", "5025": "auto/modeling_auto.py:AutoModelForTokenClassification", "5026": "auto/modeling_auto.py:AutoModelForMultipleChoice", "5027": "auto/modeling_auto.py:AutoModelForNextSentencePrediction", "5028": "auto/modeling_auto.py:AutoModelForImageClassification", "5029": "auto/modeling_auto.py:AutoModelForZeroShotImageClassification", "5030": "auto/modeling_auto.py:AutoModelForImageSegmentation", "5031": "auto/modeling_auto.py:AutoModelForSemanticSegmentation", "5032": "auto/modeling_auto.py:AutoModelForTimeSeriesPrediction", "5033": "auto/modeling_auto.py:AutoModelForUniversalSegmentation", "5034": "auto/modeling_auto.py:AutoModelForInstanceSegmentation", "5035": "auto/modeling_auto.py:AutoModelForObjectDetection", "5036": "auto/modeling_auto.py:AutoModelForZeroShotObjectDetection", "5037": "auto/modeling_auto.py:AutoModelForDepthEstimation", "5038": "auto/modeling_auto.py:AutoModelForVideoClassification", "5039": "auto/modeling_auto.py:_AutoModelForVision2Seq", "5040": "auto/modeling_auto.py:AutoModelForImageTextToText", "5041": "auto/modeling_auto.py:AutoModelForAudioClassification", "5042": "auto/modeling_auto.py:AutoModelForCTC", "5043": "auto/modeling_auto.py:AutoModelForSpeechSeq2Seq", "5044": "auto/modeling_auto.py:AutoModelForAudioFrameClassification", "5045": "auto/modeling_auto.py:AutoModelForAudioXVector", "5046": "auto/modeling_auto.py:AutoModelForTextToSpectrogram", "5047": "auto/modeling_auto.py:AutoModelForTextToWaveform", "5048": "auto/modeling_auto.py:AutoBackbone", "5049": "auto/modeling_auto.py:AutoModelForMaskedImageModeling", "5050": "auto/modeling_auto.py:AutoModelForAudioTokenization", "5051": "auto/modeling_auto.py:AutoModelWithLMHead", "5052": "auto/modeling_auto.py:AutoModelForVision2Seq", "5053": "arcee/modeling_arcee.py:ArceeMLP", "5054": "arcee/modeling_arcee.py:ArceeRMSNorm", "5055": "arcee/modeling_arcee.py:ArceeRotaryEmbedding", "5056": "arcee/modeling_arcee.py:rotate_half", "5057": "arcee/modeling_arcee.py:apply_rotary_pos_emb", "5058": "arcee/modeling_arcee.py:repeat_kv", "5059": "arcee/modeling_arcee.py:eager_attention_forward", "5060": "arcee/modeling_arcee.py:ArceeAttention", "5061": "arcee/modeling_arcee.py:ArceeDecoderLayer", "5062": "arcee/modeling_arcee.py:ArceePreTrainedModel", "5063": "arcee/modeling_arcee.py:ArceeModel", "5064": "arcee/modeling_arcee.py:ArceeForCausalLM", "5065": "arcee/modeling_arcee.py:ArceeForSequenceClassification", "5066": "arcee/modeling_arcee.py:ArceeForQuestionAnswering", "5067": "arcee/modeling_arcee.py:ArceeForTokenClassification", "5068": "poolformer/modeling_poolformer.py:drop_path", "5069": "poolformer/modeling_poolformer.py:PoolFormerDropPath", "5070": "poolformer/modeling_poolformer.py:PoolFormerEmbeddings", "5071": "poolformer/modeling_poolformer.py:PoolFormerGroupNorm", "5072": "poolformer/modeling_poolformer.py:PoolFormerPooling", "5073": "poolformer/modeling_poolformer.py:PoolFormerOutput", "5074": "poolformer/modeling_poolformer.py:PoolFormerLayer", "5075": "poolformer/modeling_poolformer.py:PoolFormerEncoder", "5076": "poolformer/modeling_poolformer.py:PoolFormerPreTrainedModel", "5077": "poolformer/modeling_poolformer.py:PoolFormerModel", "5078": "poolformer/modeling_poolformer.py:PoolFormerFinalPooler", "5079": "poolformer/modeling_poolformer.py:PoolFormerForImageClassification", "5080": "longformer/modeling_longformer.py:LongformerBaseModelOutput", "5081": "longformer/modeling_longformer.py:LongformerBaseModelOutputWithPooling", "5082": "longformer/modeling_longformer.py:LongformerMaskedLMOutput", "5083": "longformer/modeling_longformer.py:LongformerQuestionAnsweringModelOutput", "5084": "longformer/modeling_longformer.py:LongformerSequenceClassifierOutput", "5085": "longformer/modeling_longformer.py:LongformerMultipleChoiceModelOutput", "5086": "longformer/modeling_longformer.py:LongformerTokenClassifierOutput", "5087": "longformer/modeling_longformer.py:_get_question_end_index", "5088": "longformer/modeling_longformer.py:_compute_global_attention_mask", "5089": "longformer/modeling_longformer.py:create_position_ids_from_input_ids", "5090": "longformer/modeling_longformer.py:LongformerEmbeddings", "5091": "longformer/modeling_longformer.py:LongformerSelfAttention", "5092": "longformer/modeling_longformer.py:LongformerSelfOutput", "5093": "longformer/modeling_longformer.py:LongformerAttention", "5094": "longformer/modeling_longformer.py:LongformerIntermediate", "5095": "longformer/modeling_longformer.py:LongformerOutput", "5096": "longformer/modeling_longformer.py:LongformerLayer", "5097": "longformer/modeling_longformer.py:LongformerEncoder", "5098": "longformer/modeling_longformer.py:LongformerPooler", "5099": "longformer/modeling_longformer.py:LongformerLMHead", "5100": "longformer/modeling_longformer.py:LongformerPreTrainedModel", "5101": "longformer/modeling_longformer.py:LongformerModel", "5102": "longformer/modeling_longformer.py:LongformerForMaskedLM", "5103": "longformer/modeling_longformer.py:LongformerForSequenceClassification", "5104": "longformer/modeling_longformer.py:LongformerClassificationHead", "5105": "longformer/modeling_longformer.py:LongformerForQuestionAnswering", "5106": "longformer/modeling_longformer.py:LongformerForTokenClassification", "5107": "longformer/modeling_longformer.py:LongformerForMultipleChoice", "5108": "esm/modeling_esmfold.py:EsmForProteinFoldingOutput", "5109": "esm/modeling_esmfold.py:is_fp16_enabled", "5110": "esm/modeling_esmfold.py:is_deepspeed_initialized", "5111": "esm/modeling_esmfold.py:collate_dense_tensors", "5112": "esm/modeling_esmfold.py:flatten_final_dims", "5113": "esm/modeling_esmfold.py:permute_final_dims", "5114": "esm/modeling_esmfold.py:dict_multimap", "5115": "esm/modeling_esmfold.py:trunc_normal_init_", "5116": "esm/modeling_esmfold.py:ipa_point_weights_init_", "5117": "esm/modeling_esmfold.py:EsmFoldLinear", "5118": "esm/modeling_esmfold.py:EsmFoldLayerNorm", "5119": "esm/modeling_esmfold.py:softmax_no_cast", "5120": "esm/modeling_esmfold.py:EsmFoldAttention", "5121": "esm/modeling_esmfold.py:EsmFoldTriangleAttention", "5122": "esm/modeling_esmfold.py:EsmFoldTriangleMultiplicativeUpdate", "5123": "esm/modeling_esmfold.py:EsmFoldPreTrainedModel", "5124": "esm/modeling_esmfold.py:EsmFoldSelfAttention", "5125": "esm/modeling_esmfold.py:EsmFoldDropout", "5126": "esm/modeling_esmfold.py:EsmFoldSequenceToPair", "5127": "esm/modeling_esmfold.py:EsmFoldPairToSequence", "5128": "esm/modeling_esmfold.py:EsmFoldResidueMLP", "5129": "esm/modeling_esmfold.py:EsmFoldTriangularSelfAttentionBlock", "5130": "esm/modeling_esmfold.py:EsmCategoricalMixture", "5131": "esm/modeling_esmfold.py:categorical_lddt", "5132": "esm/modeling_esmfold.py:get_axial_mask", "5133": "esm/modeling_esmfold.py:EsmFoldRelativePosition", "5134": "esm/modeling_esmfold.py:EsmFoldAngleResnetBlock", "5135": "esm/modeling_esmfold.py:EsmFoldAngleResnet", "5136": "esm/modeling_esmfold.py:EsmFoldInvariantPointAttention", "5137": "esm/modeling_esmfold.py:EsmFoldBackboneUpdate", "5138": "esm/modeling_esmfold.py:EsmFoldStructureModuleTransitionLayer", "5139": "esm/modeling_esmfold.py:EsmFoldStructureModuleTransition", "5140": "esm/modeling_esmfold.py:EsmFoldStructureModule", "5141": "esm/modeling_esmfold.py:EsmFoldingTrunk", "5142": "esm/modeling_esmfold.py:EsmForProteinFolding", "5143": "esm/modeling_esm.py:rotate_half", "5144": "esm/modeling_esm.py:apply_rotary_pos_emb", "5145": "esm/modeling_esm.py:gelu", "5146": "esm/modeling_esm.py:symmetrize", "5147": "esm/modeling_esm.py:average_product_correct", "5148": "esm/modeling_esm.py:RotaryEmbedding", "5149": "esm/modeling_esm.py:EsmContactPredictionHead", "5150": "esm/modeling_esm.py:EsmEmbeddings", "5151": "esm/modeling_esm.py:eager_attention_forward", "5152": "esm/modeling_esm.py:EsmSelfAttention", "5153": "esm/modeling_esm.py:EsmSelfOutput", "5154": "esm/modeling_esm.py:EsmAttention", "5155": "esm/modeling_esm.py:EsmIntermediate", "5156": "esm/modeling_esm.py:EsmOutput", "5157": "esm/modeling_esm.py:EsmLayer", "5158": "esm/modeling_esm.py:EsmEncoder", "5159": "esm/modeling_esm.py:EsmPooler", "5160": "esm/modeling_esm.py:EsmPreTrainedModel", "5161": "esm/modeling_esm.py:EsmModel", "5162": "esm/modeling_esm.py:EsmForMaskedLM", "5163": "esm/modeling_esm.py:EsmLMHead", "5164": "esm/modeling_esm.py:EsmForSequenceClassification", "5165": "esm/modeling_esm.py:EsmForTokenClassification", "5166": "esm/modeling_esm.py:EsmClassificationHead", "5167": "esm/modeling_esm.py:create_position_ids_from_input_ids", "5168": "vilt/modeling_vilt.py:ViltForImagesAndTextClassificationOutput", "5169": "vilt/modeling_vilt.py:ViltEmbeddings", "5170": "vilt/modeling_vilt.py:TextEmbeddings", "5171": "vilt/modeling_vilt.py:ViltPatchEmbeddings", "5172": "vilt/modeling_vilt.py:ViltSelfAttention", "5173": "vilt/modeling_vilt.py:ViltSelfOutput", "5174": "vilt/modeling_vilt.py:ViltAttention", "5175": "vilt/modeling_vilt.py:ViltIntermediate", "5176": "vilt/modeling_vilt.py:ViltOutput", "5177": "vilt/modeling_vilt.py:ViltLayer", "5178": "vilt/modeling_vilt.py:ViltEncoder", "5179": "vilt/modeling_vilt.py:ViltPreTrainedModel", "5180": "vilt/modeling_vilt.py:ViltModel", "5181": "vilt/modeling_vilt.py:ViltPooler", "5182": "vilt/modeling_vilt.py:ViltForMaskedLM", "5183": "vilt/modeling_vilt.py:ViltPredictionHeadTransform", "5184": "vilt/modeling_vilt.py:ViltMLMHead", "5185": "vilt/modeling_vilt.py:ViltForQuestionAnswering", "5186": "vilt/modeling_vilt.py:ViltForImageAndTextRetrieval", "5187": "vilt/modeling_vilt.py:ViltForImagesAndTextClassification", "5188": "vilt/modeling_vilt.py:ViltForTokenClassification", "5189": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaCache", "5190": "falcon_mamba/modeling_falcon_mamba.py:_lazy_load_causal_conv1d", "5191": "falcon_mamba/modeling_falcon_mamba.py:rms_forward", "5192": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaMixer", "5193": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaRMSNorm", "5194": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaBlock", "5195": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaPreTrainedModel", "5196": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaOutput", "5197": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaCausalLMOutput", "5198": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaModel", "5199": "falcon_mamba/modeling_falcon_mamba.py:FalconMambaForCausalLM", "5200": "switch_transformers/modeling_switch_transformers.py:router_z_loss_func", "5201": "switch_transformers/modeling_switch_transformers.py:load_balancing_loss_func", "5202": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersTop1Router", "5203": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersLayerNorm", "5204": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersDenseActDense", "5205": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersSparseMLP", "5206": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersLayerFF", "5207": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersAttention", "5208": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersLayerSelfAttention", "5209": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersLayerCrossAttention", "5210": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersBlock", "5211": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersPreTrainedModel", "5212": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersStack", "5213": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersModel", "5214": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersForConditionalGeneration", "5215": "switch_transformers/modeling_switch_transformers.py:SwitchTransformersEncoderModel", "5216": "dpr/modeling_dpr.py:DPRContextEncoderOutput", "5217": "dpr/modeling_dpr.py:DPRQuestionEncoderOutput", "5218": "dpr/modeling_dpr.py:DPRReaderOutput", "5219": "dpr/modeling_dpr.py:DPRPreTrainedModel", "5220": "dpr/modeling_dpr.py:DPREncoder", "5221": "dpr/modeling_dpr.py:DPRSpanPredictor", "5222": "dpr/modeling_dpr.py:DPRPretrainedContextEncoder", "5223": "dpr/modeling_dpr.py:DPRPretrainedQuestionEncoder", "5224": "dpr/modeling_dpr.py:DPRPretrainedReader", "5225": "dpr/modeling_dpr.py:DPRContextEncoder", "5226": "dpr/modeling_dpr.py:DPRQuestionEncoder", "5227": "dpr/modeling_dpr.py:DPRReader", "5228": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2MoEGate", "5229": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2MoE", "5230": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2MLP", "5231": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2RMSNorm", "5232": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2RotaryEmbedding", "5233": "deepseek_v2/modeling_deepseek_v2.py:repeat_kv", "5234": "deepseek_v2/modeling_deepseek_v2.py:eager_attention_forward", "5235": "deepseek_v2/modeling_deepseek_v2.py:apply_rotary_emb", "5236": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2Attention", "5237": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2DecoderLayer", "5238": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2PreTrainedModel", "5239": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2Model", "5240": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2ForCausalLM", "5241": "deepseek_v2/modeling_deepseek_v2.py:DeepseekV2ForSequenceClassification", "5242": "informer/modeling_informer.py:InformerFeatureEmbedder", "5243": "informer/modeling_informer.py:InformerStdScaler", "5244": "informer/modeling_informer.py:InformerMeanScaler", "5245": "informer/modeling_informer.py:InformerNOPScaler", "5246": "informer/modeling_informer.py:InformerSinusoidalPositionalEmbedding", "5247": "informer/modeling_informer.py:InformerValueEmbedding", "5248": "informer/modeling_informer.py:InformerPreTrainedModel", "5249": "informer/modeling_informer.py:eager_attention_forward", "5250": "informer/modeling_informer.py:InformerAttention", "5251": "informer/modeling_informer.py:InformerProbSparseAttention", "5252": "informer/modeling_informer.py:InformerConvLayer", "5253": "informer/modeling_informer.py:InformerEncoderLayer", "5254": "informer/modeling_informer.py:InformerDecoderLayer", "5255": "informer/modeling_informer.py:InformerEncoder", "5256": "informer/modeling_informer.py:InformerDecoder", "5257": "informer/modeling_informer.py:InformerModel", "5258": "informer/modeling_informer.py:weighted_average", "5259": "informer/modeling_informer.py:nll", "5260": "informer/modeling_informer.py:InformerForPrediction", "5261": "camembert/modeling_camembert.py:eager_attention_forward", "5262": "camembert/modeling_camembert.py:CamembertSelfAttention", "5263": "camembert/modeling_camembert.py:CamembertCrossAttention", "5264": "camembert/modeling_camembert.py:CamembertSelfOutput", "5265": "camembert/modeling_camembert.py:CamembertAttention", "5266": "camembert/modeling_camembert.py:CamembertIntermediate", "5267": "camembert/modeling_camembert.py:CamembertOutput", "5268": "camembert/modeling_camembert.py:CamembertLayer", "5269": "camembert/modeling_camembert.py:CamembertLMHead", "5270": "camembert/modeling_camembert.py:CamembertPreTrainedModel", "5271": "camembert/modeling_camembert.py:CamembertEmbeddings", "5272": "camembert/modeling_camembert.py:CamembertEncoder", "5273": "camembert/modeling_camembert.py:CamembertPooler", "5274": "camembert/modeling_camembert.py:CamembertModel", "5275": "camembert/modeling_camembert.py:CamembertForMaskedLM", "5276": "camembert/modeling_camembert.py:CamembertClassificationHead", "5277": "camembert/modeling_camembert.py:CamembertForSequenceClassification", "5278": "camembert/modeling_camembert.py:CamembertForMultipleChoice", "5279": "camembert/modeling_camembert.py:CamembertForTokenClassification", "5280": "camembert/modeling_camembert.py:CamembertForQuestionAnswering", "5281": "camembert/modeling_camembert.py:CamembertForCausalLM", "5282": "mobilevit/modeling_mobilevit.py:make_divisible", "5283": "mobilevit/modeling_mobilevit.py:MobileViTConvLayer", "5284": "mobilevit/modeling_mobilevit.py:MobileViTInvertedResidual", "5285": "mobilevit/modeling_mobilevit.py:MobileViTMobileNetLayer", "5286": "mobilevit/modeling_mobilevit.py:MobileViTSelfAttention", "5287": "mobilevit/modeling_mobilevit.py:MobileViTSelfOutput", "5288": "mobilevit/modeling_mobilevit.py:MobileViTAttention", "5289": "mobilevit/modeling_mobilevit.py:MobileViTIntermediate", "5290": "mobilevit/modeling_mobilevit.py:MobileViTOutput", "5291": "mobilevit/modeling_mobilevit.py:MobileViTTransformerLayer", "5292": "mobilevit/modeling_mobilevit.py:MobileViTTransformer", "5293": "mobilevit/modeling_mobilevit.py:MobileViTLayer", "5294": "mobilevit/modeling_mobilevit.py:MobileViTEncoder", "5295": "mobilevit/modeling_mobilevit.py:MobileViTPreTrainedModel", "5296": "mobilevit/modeling_mobilevit.py:MobileViTModel", "5297": "mobilevit/modeling_mobilevit.py:MobileViTForImageClassification", "5298": "mobilevit/modeling_mobilevit.py:MobileViTASPPPooling", "5299": "mobilevit/modeling_mobilevit.py:MobileViTASPP", "5300": "mobilevit/modeling_mobilevit.py:MobileViTDeepLabV3", "5301": "mobilevit/modeling_mobilevit.py:MobileViTForSemanticSegmentation", "5302": "albert/modeling_albert.py:AlbertEmbeddings", "5303": "albert/modeling_albert.py:eager_attention_forward", "5304": "albert/modeling_albert.py:AlbertAttention", "5305": "albert/modeling_albert.py:AlbertLayer", "5306": "albert/modeling_albert.py:AlbertLayerGroup", "5307": "albert/modeling_albert.py:AlbertTransformer", "5308": "albert/modeling_albert.py:AlbertPreTrainedModel", "5309": "albert/modeling_albert.py:AlbertForPreTrainingOutput", "5310": "albert/modeling_albert.py:AlbertModel", "5311": "albert/modeling_albert.py:AlbertForPreTraining", "5312": "albert/modeling_albert.py:AlbertMLMHead", "5313": "albert/modeling_albert.py:AlbertSOPHead", "5314": "albert/modeling_albert.py:AlbertForMaskedLM", "5315": "albert/modeling_albert.py:AlbertForSequenceClassification", "5316": "albert/modeling_albert.py:AlbertForTokenClassification", "5317": "albert/modeling_albert.py:AlbertForQuestionAnswering", "5318": "albert/modeling_albert.py:AlbertForMultipleChoice", "5319": "bert_generation/modeling_bert_generation.py:BertGenerationSelfOutput", "5320": "bert_generation/modeling_bert_generation.py:eager_attention_forward", "5321": "bert_generation/modeling_bert_generation.py:BertGenerationSelfAttention", "5322": "bert_generation/modeling_bert_generation.py:BertGenerationCrossAttention", "5323": "bert_generation/modeling_bert_generation.py:BertGenerationAttention", "5324": "bert_generation/modeling_bert_generation.py:BertGenerationIntermediate", "5325": "bert_generation/modeling_bert_generation.py:BertGenerationOutput", "5326": "bert_generation/modeling_bert_generation.py:BertGenerationLayer", "5327": "bert_generation/modeling_bert_generation.py:BertEncoder", "5328": "bert_generation/modeling_bert_generation.py:BertGenerationEmbeddings", "5329": "bert_generation/modeling_bert_generation.py:BertGenerationPreTrainedModel", "5330": "bert_generation/modeling_bert_generation.py:BertGenerationEncoder", "5331": "bert_generation/modeling_bert_generation.py:BertGenerationOnlyLMHead", "5332": "bert_generation/modeling_bert_generation.py:BertGenerationDecoder", "5333": "swiftformer/modeling_swiftformer.py:SwiftFormerPatchEmbedding", "5334": "swiftformer/modeling_swiftformer.py:drop_path", "5335": "swiftformer/modeling_swiftformer.py:SwiftFormerDropPath", "5336": "swiftformer/modeling_swiftformer.py:SwiftFormerEmbeddings", "5337": "swiftformer/modeling_swiftformer.py:SwiftFormerConvEncoder", "5338": "swiftformer/modeling_swiftformer.py:SwiftFormerMlp", "5339": "swiftformer/modeling_swiftformer.py:SwiftFormerEfficientAdditiveAttention", "5340": "swiftformer/modeling_swiftformer.py:SwiftFormerLocalRepresentation", "5341": "swiftformer/modeling_swiftformer.py:SwiftFormerEncoderBlock", "5342": "swiftformer/modeling_swiftformer.py:SwiftFormerStage", "5343": "swiftformer/modeling_swiftformer.py:SwiftFormerEncoder", "5344": "swiftformer/modeling_swiftformer.py:SwiftFormerPreTrainedModel", "5345": "swiftformer/modeling_swiftformer.py:SwiftFormerModel", "5346": "swiftformer/modeling_swiftformer.py:SwiftFormerForImageClassification", "5347": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesFeatureEmbedder", "5348": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesStdScaler", "5349": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesMeanScaler", "5350": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesNOPScaler", "5351": "time_series_transformer/modeling_time_series_transformer.py:nll", "5352": "time_series_transformer/modeling_time_series_transformer.py:weighted_average", "5353": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesSinusoidalPositionalEmbedding", "5354": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesValueEmbedding", "5355": "time_series_transformer/modeling_time_series_transformer.py:eager_attention_forward", "5356": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerAttention", "5357": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerEncoderLayer", "5358": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerDecoderLayer", "5359": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerPreTrainedModel", "5360": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerEncoder", "5361": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerDecoder", "5362": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerModel", "5363": "time_series_transformer/modeling_time_series_transformer.py:TimeSeriesTransformerForPrediction", "5364": "bart/modeling_bart.py:shift_tokens_right", "5365": "bart/modeling_bart.py:BartLearnedPositionalEmbedding", "5366": "bart/modeling_bart.py:BartScaledWordEmbedding", "5367": "bart/modeling_bart.py:eager_attention_forward", "5368": "bart/modeling_bart.py:BartAttention", "5369": "bart/modeling_bart.py:BartEncoderLayer", "5370": "bart/modeling_bart.py:BartDecoderLayer", "5371": "bart/modeling_bart.py:BartClassificationHead", "5372": "bart/modeling_bart.py:BartPreTrainedModel", "5373": "bart/modeling_bart.py:PretrainedBartModel", "5374": "bart/modeling_bart.py:BartPretrainedModel", "5375": "bart/modeling_bart.py:BartEncoder", "5376": "bart/modeling_bart.py:BartDecoder", "5377": "bart/modeling_bart.py:BartModel", "5378": "bart/modeling_bart.py:BartForConditionalGeneration", "5379": "bart/modeling_bart.py:BartForSequenceClassification", "5380": "bart/modeling_bart.py:BartForQuestionAnswering", "5381": "bart/modeling_bart.py:BartDecoderWrapper", "5382": "bart/modeling_bart.py:BartForCausalLM", "5383": "tvp/modeling_tvp.py:TvpVideoGroundingOutput", "5384": "tvp/modeling_tvp.py:TvpLoss", "5385": "tvp/modeling_tvp.py:TvpVisionModel", "5386": "tvp/modeling_tvp.py:TvpVisualInputEmbedding", "5387": "tvp/modeling_tvp.py:TvpTextInputEmbeddings", "5388": "tvp/modeling_tvp.py:TvpAttention", "5389": "tvp/modeling_tvp.py:TvpIntermediate", "5390": "tvp/modeling_tvp.py:TvpOutputLayer", "5391": "tvp/modeling_tvp.py:TvpEncodeLayer", "5392": "tvp/modeling_tvp.py:TvpEncoder", "5393": "tvp/modeling_tvp.py:TvpPooler", "5394": "tvp/modeling_tvp.py:TvpPreTrainedModel", "5395": "tvp/modeling_tvp.py:TvpFrameDownPadPrompter", "5396": "tvp/modeling_tvp.py:TvpFramePadPrompter", "5397": "tvp/modeling_tvp.py:TvpModel", "5398": "tvp/modeling_tvp.py:TvpVideoGroundingHead", "5399": "tvp/modeling_tvp.py:TvpForVideoGrounding", "5400": "colqwen2/modeling_colqwen2.py:ColQwen2PreTrainedModel", "5401": "colqwen2/modeling_colqwen2.py:ColQwen2ForRetrievalOutput", "5402": "colqwen2/modeling_colqwen2.py:ColQwen2ForRetrieval", "5403": "bridgetower/modeling_bridgetower.py:BridgeTowerModelOutput", "5404": "bridgetower/modeling_bridgetower.py:BridgeTowerContrastiveOutput", "5405": "bridgetower/modeling_bridgetower.py:BridgeTowerResidualAttention", "5406": "bridgetower/modeling_bridgetower.py:BridgeTowerTransformer", "5407": "bridgetower/modeling_bridgetower.py:BridgeTowerVisionEmbeddings", "5408": "bridgetower/modeling_bridgetower.py:BridgeTowerVisionTransformer", "5409": "bridgetower/modeling_bridgetower.py:BridgeTowerLinkTower", "5410": "bridgetower/modeling_bridgetower.py:BridgeTowerSelfOutput", "5411": "bridgetower/modeling_bridgetower.py:BridgeTowerIntermediate", "5412": "bridgetower/modeling_bridgetower.py:BridgeTowerOutput", "5413": "bridgetower/modeling_bridgetower.py:BridgeTowerPooler", "5414": "bridgetower/modeling_bridgetower.py:eager_attention_forward", "5415": "bridgetower/modeling_bridgetower.py:BridgeTowerSelfAttention", "5416": "bridgetower/modeling_bridgetower.py:BridgeTowerCrossAttention", "5417": "bridgetower/modeling_bridgetower.py:BridgeTowerAttention", "5418": "bridgetower/modeling_bridgetower.py:BridgeTowerBertCrossLayer", "5419": "bridgetower/modeling_bridgetower.py:BridgeTowerTextLayer", "5420": "bridgetower/modeling_bridgetower.py:BridgeTowerTextEncoder", "5421": "bridgetower/modeling_bridgetower.py:BridgeTowerTextEmbeddings", "5422": "bridgetower/modeling_bridgetower.py:BridgeTowerPreTrainedModel", "5423": "bridgetower/modeling_bridgetower.py:BridgeTowerVisionModel", "5424": "bridgetower/modeling_bridgetower.py:BridgeTowerTextModel", "5425": "bridgetower/modeling_bridgetower.py:BridgeTowerModel", "5426": "bridgetower/modeling_bridgetower.py:BridgeTowerPredictionHeadTransform", "5427": "bridgetower/modeling_bridgetower.py:BridgeTowerMLMHead", "5428": "bridgetower/modeling_bridgetower.py:BridgeTowerITMHead", "5429": "bridgetower/modeling_bridgetower.py:BridgeTowerForMaskedLM", "5430": "bridgetower/modeling_bridgetower.py:BridgeTowerForImageAndTextRetrieval", "5431": "bridgetower/modeling_bridgetower.py:BridgeTowerContrastiveHead", "5432": "bridgetower/modeling_bridgetower.py:BridgeTowerForContrastiveLearning", "5433": "autoformer/modeling_autoformer.py:AutoFormerDecoderOutput", "5434": "autoformer/modeling_autoformer.py:AutoformerModelOutput", "5435": "autoformer/modeling_autoformer.py:AutoformerFeatureEmbedder", "5436": "autoformer/modeling_autoformer.py:AutoformerStdScaler", "5437": "autoformer/modeling_autoformer.py:AutoformerMeanScaler", "5438": "autoformer/modeling_autoformer.py:AutoformerNOPScaler", "5439": "autoformer/modeling_autoformer.py:weighted_average", "5440": "autoformer/modeling_autoformer.py:nll", "5441": "autoformer/modeling_autoformer.py:AutoformerSinusoidalPositionalEmbedding", "5442": "autoformer/modeling_autoformer.py:AutoformerValueEmbedding", "5443": "autoformer/modeling_autoformer.py:AutoformerSeriesDecompositionLayer", "5444": "autoformer/modeling_autoformer.py:AutoformerLayernorm", "5445": "autoformer/modeling_autoformer.py:AutoformerAttention", "5446": "autoformer/modeling_autoformer.py:AutoformerEncoderLayer", "5447": "autoformer/modeling_autoformer.py:AutoformerDecoderLayer", "5448": "autoformer/modeling_autoformer.py:AutoformerPreTrainedModel", "5449": "autoformer/modeling_autoformer.py:AutoformerEncoder", "5450": "autoformer/modeling_autoformer.py:AutoformerDecoder", "5451": "autoformer/modeling_autoformer.py:AutoformerModel", "5452": "autoformer/modeling_autoformer.py:AutoformerForPrediction", "5453": "granitemoehybrid/modeling_granitemoehybrid.py:rotate_half", "5454": "granitemoehybrid/modeling_granitemoehybrid.py:apply_rotary_pos_emb", "5455": "granitemoehybrid/modeling_granitemoehybrid.py:repeat_kv", "5456": "granitemoehybrid/modeling_granitemoehybrid.py:eager_attention_forward", "5457": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridAttention", "5458": "granitemoehybrid/modeling_granitemoehybrid.py:HybridMambaAttentionDynamicCache", "5459": "granitemoehybrid/modeling_granitemoehybrid.py:pad_tensor_by_size", "5460": "granitemoehybrid/modeling_granitemoehybrid.py:reshape_into_chunks", "5461": "granitemoehybrid/modeling_granitemoehybrid.py:segment_sum", "5462": "granitemoehybrid/modeling_granitemoehybrid.py:apply_mask_to_padding_states", "5463": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridMambaLayer", "5464": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridRMSNormGated", "5465": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridMLP", "5466": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteFlashAttentionKwargs", "5467": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridRMSNorm", "5468": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridParallelExperts", "5469": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridTopKGating", "5470": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridMoE", "5471": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridDecoderLayer", "5472": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridPreTrainedModel", "5473": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridRotaryEmbedding", "5474": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridModel", "5475": "granitemoehybrid/modeling_granitemoehybrid.py:load_balancing_loss_func", "5476": "granitemoehybrid/modeling_granitemoehybrid.py:GraniteMoeHybridForCausalLM", "5477": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLModelOutputWithPast", "5478": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLCausalLMOutputWithPast", "5479": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLRotaryEmbedding", "5480": "qwen2_vl/modeling_qwen2_vl.py:rotate_half", "5481": "qwen2_vl/modeling_qwen2_vl.py:apply_multimodal_rotary_pos_emb", "5482": "qwen2_vl/modeling_qwen2_vl.py:apply_rotary_pos_emb_vision", "5483": "qwen2_vl/modeling_qwen2_vl.py:VisionRotaryEmbedding", "5484": "qwen2_vl/modeling_qwen2_vl.py:PatchEmbed", "5485": "qwen2_vl/modeling_qwen2_vl.py:PatchMerger", "5486": "qwen2_vl/modeling_qwen2_vl.py:VisionMlp", "5487": "qwen2_vl/modeling_qwen2_vl.py:repeat_kv", "5488": "qwen2_vl/modeling_qwen2_vl.py:eager_attention_forward", "5489": "qwen2_vl/modeling_qwen2_vl.py:VisionAttention", "5490": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLVisionBlock", "5491": "qwen2_vl/modeling_qwen2_vl.py:Qwen2MLP", "5492": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLAttention", "5493": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLDecoderLayer", "5494": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLPreTrainedModel", "5495": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VisionTransformerPretrainedModel", "5496": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLTextModel", "5497": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLModel", "5498": "qwen2_vl/modeling_qwen2_vl.py:Qwen2VLForConditionalGeneration", "5499": "dbrx/modeling_dbrx.py:DbrxRotaryEmbedding", "5500": "dbrx/modeling_dbrx.py:rotate_half", "5501": "dbrx/modeling_dbrx.py:apply_rotary_pos_emb", "5502": "dbrx/modeling_dbrx.py:repeat_kv", "5503": "dbrx/modeling_dbrx.py:load_balancing_loss_func", "5504": "dbrx/modeling_dbrx.py:DbrxAttention", "5505": "dbrx/modeling_dbrx.py:DbrxFlashAttention2", "5506": "dbrx/modeling_dbrx.py:DbrxSdpaAttention", "5507": "dbrx/modeling_dbrx.py:DbrxNormAttentionNorm", "5508": "dbrx/modeling_dbrx.py:DbrxRouter", "5509": "dbrx/modeling_dbrx.py:DbrxExpertGLU", "5510": "dbrx/modeling_dbrx.py:DbrxExperts", "5511": "dbrx/modeling_dbrx.py:DbrxFFN", "5512": "dbrx/modeling_dbrx.py:DbrxBlock", "5513": "dbrx/modeling_dbrx.py:DbrxPreTrainedModel", "5514": "dbrx/modeling_dbrx.py:DbrxModel", "5515": "dbrx/modeling_dbrx.py:DbrxForCausalLM", "5516": "deberta/modeling_deberta.py:DebertaLayerNorm", "5517": "deberta/modeling_deberta.py:DebertaSelfOutput", "5518": "deberta/modeling_deberta.py:build_relative_position", "5519": "deberta/modeling_deberta.py:c2p_dynamic_expand", "5520": "deberta/modeling_deberta.py:p2c_dynamic_expand", "5521": "deberta/modeling_deberta.py:pos_dynamic_expand", "5522": "deberta/modeling_deberta.py:scaled_size_sqrt", "5523": "deberta/modeling_deberta.py:build_rpos", "5524": "deberta/modeling_deberta.py:compute_attention_span", "5525": "deberta/modeling_deberta.py:uneven_size_corrected", "5526": "deberta/modeling_deberta.py:DisentangledSelfAttention", "5527": "deberta/modeling_deberta.py:DebertaEmbeddings", "5528": "deberta/modeling_deberta.py:DebertaAttention", "5529": "deberta/modeling_deberta.py:DebertaIntermediate", "5530": "deberta/modeling_deberta.py:DebertaOutput", "5531": "deberta/modeling_deberta.py:DebertaLayer", "5532": "deberta/modeling_deberta.py:DebertaEncoder", "5533": "deberta/modeling_deberta.py:DebertaPreTrainedModel", "5534": "deberta/modeling_deberta.py:DebertaModel", "5535": "deberta/modeling_deberta.py:LegacyDebertaPredictionHeadTransform", "5536": "deberta/modeling_deberta.py:LegacyDebertaLMPredictionHead", "5537": "deberta/modeling_deberta.py:LegacyDebertaOnlyMLMHead", "5538": "deberta/modeling_deberta.py:DebertaLMPredictionHead", "5539": "deberta/modeling_deberta.py:DebertaOnlyMLMHead", "5540": "deberta/modeling_deberta.py:DebertaForMaskedLM", "5541": "deberta/modeling_deberta.py:ContextPooler", "5542": "deberta/modeling_deberta.py:DebertaForSequenceClassification", "5543": "deberta/modeling_deberta.py:DebertaForTokenClassification", "5544": "deberta/modeling_deberta.py:DebertaForQuestionAnswering", "5545": "cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionMultiModalProjector", "5546": "cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionModelOutputWithPast", "5547": "cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionCausalLMOutputWithPast", "5548": "cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionPreTrainedModel", "5549": "cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionModel", "5550": "cohere2_vision/modeling_cohere2_vision.py:Cohere2VisionForConditionalGeneration", "5551": "plbart/modeling_plbart.py:PLBartScaledWordEmbedding", "5552": "plbart/modeling_plbart.py:PLBartPreTrainedModel", "5553": "plbart/modeling_plbart.py:PLBartLearnedPositionalEmbedding", "5554": "plbart/modeling_plbart.py:eager_attention_forward", "5555": "plbart/modeling_plbart.py:PLBartAttention", "5556": "plbart/modeling_plbart.py:PLBartEncoderLayer", "5557": "plbart/modeling_plbart.py:PLBartEncoder", "5558": "plbart/modeling_plbart.py:PLBartDecoderLayer", "5559": "plbart/modeling_plbart.py:PLBartDecoder", "5560": "plbart/modeling_plbart.py:shift_tokens_right", "5561": "plbart/modeling_plbart.py:PLBartModel", "5562": "plbart/modeling_plbart.py:PLBartForConditionalGeneration", "5563": "plbart/modeling_plbart.py:PLBartClassificationHead", "5564": "plbart/modeling_plbart.py:PLBartForSequenceClassification", "5565": "plbart/modeling_plbart.py:PLBartDecoderWrapper", "5566": "plbart/modeling_plbart.py:PLBartForCausalLM", "5567": "layoutlm/modeling_layoutlm.py:LayoutLMEmbeddings", "5568": "layoutlm/modeling_layoutlm.py:eager_attention_forward", "5569": "layoutlm/modeling_layoutlm.py:LayoutLMSelfAttention", "5570": "layoutlm/modeling_layoutlm.py:LayoutLMSelfOutput", "5571": "layoutlm/modeling_layoutlm.py:LayoutLMAttention", "5572": "layoutlm/modeling_layoutlm.py:LayoutLMIntermediate", "5573": "layoutlm/modeling_layoutlm.py:LayoutLMOutput", "5574": "layoutlm/modeling_layoutlm.py:LayoutLMLayer", "5575": "layoutlm/modeling_layoutlm.py:LayoutLMEncoder", "5576": "layoutlm/modeling_layoutlm.py:LayoutLMPooler", "5577": "layoutlm/modeling_layoutlm.py:LayoutLMPredictionHeadTransform", "5578": "layoutlm/modeling_layoutlm.py:LayoutLMLMPredictionHead", "5579": "layoutlm/modeling_layoutlm.py:LayoutLMOnlyMLMHead", "5580": "layoutlm/modeling_layoutlm.py:LayoutLMPreTrainedModel", "5581": "layoutlm/modeling_layoutlm.py:LayoutLMModel", "5582": "layoutlm/modeling_layoutlm.py:LayoutLMForMaskedLM", "5583": "layoutlm/modeling_layoutlm.py:LayoutLMForSequenceClassification", "5584": "layoutlm/modeling_layoutlm.py:LayoutLMForTokenClassification", "5585": "layoutlm/modeling_layoutlm.py:LayoutLMForQuestionAnswering", "5586": "clvp/modeling_clvp.py:contrastive_loss", "5587": "clvp/modeling_clvp.py:clvp_loss", "5588": "clvp/modeling_clvp.py:rotate_half", "5589": "clvp/modeling_clvp.py:apply_rotary_pos_emb", "5590": "clvp/modeling_clvp.py:_pad_extra_bos_eos_tokens", "5591": "clvp/modeling_clvp.py:ClvpEncoderOutput", "5592": "clvp/modeling_clvp.py:ClvpOutput", "5593": "clvp/modeling_clvp.py:ClvpRMSNorm", "5594": "clvp/modeling_clvp.py:ClvpRotaryPositionalEmbedding", "5595": "clvp/modeling_clvp.py:ClvpSelfAttention", "5596": "clvp/modeling_clvp.py:ClvpGatedLinearUnit", "5597": "clvp/modeling_clvp.py:ClvpEncoderMLP", "5598": "clvp/modeling_clvp.py:ClvpEncoderLayer", "5599": "clvp/modeling_clvp.py:ClvpSequenceSummary", "5600": "clvp/modeling_clvp.py:ClvpDecoderMLP", "5601": "clvp/modeling_clvp.py:ClvpDecoderLayer", "5602": "clvp/modeling_clvp.py:ClvpConditioningEncoder", "5603": "clvp/modeling_clvp.py:ClvpPreTrainedModel", "5604": "clvp/modeling_clvp.py:ClvpEncoder", "5605": "clvp/modeling_clvp.py:ClvpDecoder", "5606": "clvp/modeling_clvp.py:ClvpModel", "5607": "clvp/modeling_clvp.py:ClvpForCausalLM", "5608": "clvp/modeling_clvp.py:ClvpModelForConditionalGeneration", "5609": "qwen3_moe/modeling_qwen3_moe.py:rotate_half", "5610": "qwen3_moe/modeling_qwen3_moe.py:apply_rotary_pos_emb", "5611": "qwen3_moe/modeling_qwen3_moe.py:repeat_kv", "5612": "qwen3_moe/modeling_qwen3_moe.py:eager_attention_forward", "5613": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeAttention", "5614": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeMLP", "5615": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeSparseMoeBlock", "5616": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeRMSNorm", "5617": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeDecoderLayer", "5618": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeRotaryEmbedding", "5619": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoePreTrainedModel", "5620": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeModel", "5621": "qwen3_moe/modeling_qwen3_moe.py:load_balancing_loss_func", "5622": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeForCausalLM", "5623": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeForSequenceClassification", "5624": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeForTokenClassification", "5625": "qwen3_moe/modeling_qwen3_moe.py:Qwen3MoeForQuestionAnswering", "5626": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTEmbeddings", "5627": "dinov3_vit/modeling_dinov3_vit.py:get_patches_center_coordinates", "5628": "dinov3_vit/modeling_dinov3_vit.py:augment_patches_center_coordinates", "5629": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTRopePositionEmbedding", "5630": "dinov3_vit/modeling_dinov3_vit.py:rotate_half", "5631": "dinov3_vit/modeling_dinov3_vit.py:eager_attention_forward", "5632": "dinov3_vit/modeling_dinov3_vit.py:apply_rotary_pos_emb", "5633": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTAttention", "5634": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTLayerScale", "5635": "dinov3_vit/modeling_dinov3_vit.py:drop_path", "5636": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTDropPath", "5637": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTMLP", "5638": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTGatedMLP", "5639": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTLayer", "5640": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTPreTrainedModel", "5641": "dinov3_vit/modeling_dinov3_vit.py:DINOv3ViTModel", "5642": "pvt/modeling_pvt.py:drop_path", "5643": "pvt/modeling_pvt.py:PvtDropPath", "5644": "pvt/modeling_pvt.py:PvtPatchEmbeddings", "5645": "pvt/modeling_pvt.py:PvtSelfOutput", "5646": "pvt/modeling_pvt.py:PvtEfficientSelfAttention", "5647": "pvt/modeling_pvt.py:PvtAttention", "5648": "pvt/modeling_pvt.py:PvtFFN", "5649": "pvt/modeling_pvt.py:PvtLayer", "5650": "pvt/modeling_pvt.py:PvtEncoder", "5651": "pvt/modeling_pvt.py:PvtPreTrainedModel", "5652": "pvt/modeling_pvt.py:PvtModel", "5653": "pvt/modeling_pvt.py:PvtForImageClassification", "5654": "tapas/modeling_tapas.py:TableQuestionAnsweringOutput", "5655": "tapas/modeling_tapas.py:TapasEmbeddings", "5656": "tapas/modeling_tapas.py:TapasSelfAttention", "5657": "tapas/modeling_tapas.py:TapasSelfOutput", "5658": "tapas/modeling_tapas.py:TapasAttention", "5659": "tapas/modeling_tapas.py:TapasIntermediate", "5660": "tapas/modeling_tapas.py:TapasOutput", "5661": "tapas/modeling_tapas.py:TapasLayer", "5662": "tapas/modeling_tapas.py:TapasEncoder", "5663": "tapas/modeling_tapas.py:TapasPooler", "5664": "tapas/modeling_tapas.py:TapasPredictionHeadTransform", "5665": "tapas/modeling_tapas.py:TapasLMPredictionHead", "5666": "tapas/modeling_tapas.py:TapasOnlyMLMHead", "5667": "tapas/modeling_tapas.py:TapasPreTrainedModel", "5668": "tapas/modeling_tapas.py:TapasModel", "5669": "tapas/modeling_tapas.py:TapasForMaskedLM", "5670": "tapas/modeling_tapas.py:TapasForQuestionAnswering", "5671": "tapas/modeling_tapas.py:TapasForSequenceClassification", "5672": "tapas/modeling_tapas.py:AverageApproximationFunction", "5673": "tapas/modeling_tapas.py:IndexMap", "5674": "tapas/modeling_tapas.py:ProductIndexMap", "5675": "tapas/modeling_tapas.py:gather", "5676": "tapas/modeling_tapas.py:flatten", "5677": "tapas/modeling_tapas.py:range_index_map", "5678": "tapas/modeling_tapas.py:_segment_reduce", "5679": "tapas/modeling_tapas.py:reduce_sum", "5680": "tapas/modeling_tapas.py:reduce_mean", "5681": "tapas/modeling_tapas.py:reduce_max", "5682": "tapas/modeling_tapas.py:reduce_min", "5683": "tapas/modeling_tapas.py:compute_column_logits", "5684": "tapas/modeling_tapas.py:_single_column_cell_selection_loss", "5685": "tapas/modeling_tapas.py:compute_token_logits", "5686": "tapas/modeling_tapas.py:_calculate_aggregate_mask", "5687": "tapas/modeling_tapas.py:_calculate_aggregation_loss_known", "5688": "tapas/modeling_tapas.py:_calculate_aggregation_loss_unknown", "5689": "tapas/modeling_tapas.py:_calculate_aggregation_loss", "5690": "tapas/modeling_tapas.py:_calculate_expected_result", "5691": "tapas/modeling_tapas.py:huber_loss", "5692": "tapas/modeling_tapas.py:_calculate_regression_loss", "5693": "visual_bert/modeling_visual_bert.py:VisualBertEmbeddings", "5694": "visual_bert/modeling_visual_bert.py:VisualBertSelfAttention", "5695": "visual_bert/modeling_visual_bert.py:VisualBertSelfOutput", "5696": "visual_bert/modeling_visual_bert.py:VisualBertAttention", "5697": "visual_bert/modeling_visual_bert.py:VisualBertIntermediate", "5698": "visual_bert/modeling_visual_bert.py:VisualBertOutput", "5699": "visual_bert/modeling_visual_bert.py:VisualBertLayer", "5700": "visual_bert/modeling_visual_bert.py:VisualBertEncoder", "5701": "visual_bert/modeling_visual_bert.py:VisualBertPooler", "5702": "visual_bert/modeling_visual_bert.py:VisualBertPredictionHeadTransform", "5703": "visual_bert/modeling_visual_bert.py:VisualBertLMPredictionHead", "5704": "visual_bert/modeling_visual_bert.py:VisualBertPreTrainingHeads", "5705": "visual_bert/modeling_visual_bert.py:VisualBertPreTrainedModel", "5706": "visual_bert/modeling_visual_bert.py:VisualBertForPreTrainingOutput", "5707": "visual_bert/modeling_visual_bert.py:VisualBertModel", "5708": "visual_bert/modeling_visual_bert.py:VisualBertForPreTraining", "5709": "visual_bert/modeling_visual_bert.py:VisualBertForMultipleChoice", "5710": "visual_bert/modeling_visual_bert.py:VisualBertForQuestionAnswering", "5711": "visual_bert/modeling_visual_bert.py:VisualBertForVisualReasoning", "5712": "visual_bert/modeling_visual_bert.py:VisualBertRegionToPhraseAttention", "5713": "visual_bert/modeling_visual_bert.py:VisualBertForRegionToPhraseAlignment", "5714": "internvl/modeling_internvl.py:InternVLVisionRMSNorm", "5715": "internvl/modeling_internvl.py:eager_attention_forward", "5716": "internvl/modeling_internvl.py:InternVLVisionAttention", "5717": "internvl/modeling_internvl.py:InternVLVisionModelOutputWithPooling", "5718": "internvl/modeling_internvl.py:InternVLVisionPatchEmbeddings", "5719": "internvl/modeling_internvl.py:InternVLVisionEmbeddings", "5720": "internvl/modeling_internvl.py:InternVLVisionMLP", "5721": "internvl/modeling_internvl.py:InternVLVisionLayer", "5722": "internvl/modeling_internvl.py:InternVLVisionEncoder", "5723": "internvl/modeling_internvl.py:InternVLVisionPreTrainedModel", "5724": "internvl/modeling_internvl.py:InternVLVisionModel", "5725": "internvl/modeling_internvl.py:InternVLPreTrainedModel", "5726": "internvl/modeling_internvl.py:InternVLMultiModalProjector", "5727": "internvl/modeling_internvl.py:InternVLModelOutputWithPast", "5728": "internvl/modeling_internvl.py:InternVLModel", "5729": "internvl/modeling_internvl.py:InternVLCausalLMOutputWithPast", "5730": "internvl/modeling_internvl.py:InternVLForConditionalGeneration", "5731": "codegen/modeling_codegen.py:create_sinusoidal_positions", "5732": "codegen/modeling_codegen.py:rotate_every_two", "5733": "codegen/modeling_codegen.py:apply_rotary_pos_emb", "5734": "codegen/modeling_codegen.py:CodeGenAttention", "5735": "codegen/modeling_codegen.py:CodeGenMLP", "5736": "codegen/modeling_codegen.py:CodeGenBlock", "5737": "codegen/modeling_codegen.py:CodeGenPreTrainedModel", "5738": "codegen/modeling_codegen.py:CodeGenModel", "5739": "codegen/modeling_codegen.py:CodeGenForCausalLM", "5740": "ernie4_5/modeling_ernie4_5.py:Ernie4_5RotaryEmbedding", "5741": "ernie4_5/modeling_ernie4_5.py:Ernie4_5MLP", "5742": "ernie4_5/modeling_ernie4_5.py:rotate_half", "5743": "ernie4_5/modeling_ernie4_5.py:repeat_kv", "5744": "ernie4_5/modeling_ernie4_5.py:eager_attention_forward", "5745": "ernie4_5/modeling_ernie4_5.py:apply_rotary_pos_emb", "5746": "ernie4_5/modeling_ernie4_5.py:Ernie4_5Attention", "5747": "ernie4_5/modeling_ernie4_5.py:Ernie4_5RMSNorm", "5748": "ernie4_5/modeling_ernie4_5.py:Ernie4_5DecoderLayer", "5749": "ernie4_5/modeling_ernie4_5.py:Ernie4_5PreTrainedModel", "5750": "ernie4_5/modeling_ernie4_5.py:Ernie4_5Model", "5751": "ernie4_5/modeling_ernie4_5.py:Ernie4_5ForCausalLM", "5752": "eomt/modeling_eomt.py:EomtForUniversalSegmentationOutput", "5753": "eomt/modeling_eomt.py:sample_point", "5754": "eomt/modeling_eomt.py:pair_wise_dice_loss", "5755": "eomt/modeling_eomt.py:pair_wise_sigmoid_cross_entropy_loss", "5756": "eomt/modeling_eomt.py:EomtHungarianMatcher", "5757": "eomt/modeling_eomt.py:dice_loss", "5758": "eomt/modeling_eomt.py:sigmoid_cross_entropy_loss", "5759": "eomt/modeling_eomt.py:EomtLoss", "5760": "eomt/modeling_eomt.py:EomtPatchEmbeddings", "5761": "eomt/modeling_eomt.py:EomtEmbeddings", "5762": "eomt/modeling_eomt.py:eager_attention_forward", "5763": "eomt/modeling_eomt.py:EomtAttention", "5764": "eomt/modeling_eomt.py:EomtLayerScale", "5765": "eomt/modeling_eomt.py:drop_path", "5766": "eomt/modeling_eomt.py:EomtDropPath", "5767": "eomt/modeling_eomt.py:EomtMLP", "5768": "eomt/modeling_eomt.py:EomtSwiGLUFFN", "5769": "eomt/modeling_eomt.py:EomtLayer", "5770": "eomt/modeling_eomt.py:EomtLayerNorm2d", "5771": "eomt/modeling_eomt.py:EomtScaleLayer", "5772": "eomt/modeling_eomt.py:EomtScaleBlock", "5773": "eomt/modeling_eomt.py:EomtMaskHead", "5774": "eomt/modeling_eomt.py:EomtPreTrainedModel", "5775": "eomt/modeling_eomt.py:EomtForUniversalSegmentation", "5776": "parakeet/modeling_parakeet.py:ParakeetEncoderRelPositionalEncoding", "5777": "parakeet/modeling_parakeet.py:ParakeetEncoderFeedForward", "5778": "parakeet/modeling_parakeet.py:ParakeetEncoderConvolutionModule", "5779": "parakeet/modeling_parakeet.py:repeat_kv", "5780": "parakeet/modeling_parakeet.py:eager_attention_forward", "5781": "parakeet/modeling_parakeet.py:ParakeetEncoderAttention", "5782": "parakeet/modeling_parakeet.py:ParakeetEncoderSubsamplingConv2D", "5783": "parakeet/modeling_parakeet.py:ParakeetEncoderBlock", "5784": "parakeet/modeling_parakeet.py:ParakeetPreTrainedModel", "5785": "parakeet/modeling_parakeet.py:ParakeetEncoder", "5786": "parakeet/modeling_parakeet.py:ParakeetGenerateOutput", "5787": "parakeet/modeling_parakeet.py:ParakeetForCTC", "5788": "seggpt/modeling_seggpt.py:SegGptEncoderOutput", "5789": "seggpt/modeling_seggpt.py:SegGptImageSegmentationOutput", "5790": "seggpt/modeling_seggpt.py:SegGptPatchEmbeddings", "5791": "seggpt/modeling_seggpt.py:SegGptEmbeddings", "5792": "seggpt/modeling_seggpt.py:SegGptAttention", "5793": "seggpt/modeling_seggpt.py:SegGptMlp", "5794": "seggpt/modeling_seggpt.py:drop_path", "5795": "seggpt/modeling_seggpt.py:SegGptDropPath", "5796": "seggpt/modeling_seggpt.py:SegGptLayer", "5797": "seggpt/modeling_seggpt.py:SegGptEncoder", "5798": "seggpt/modeling_seggpt.py:SegGptLayerNorm", "5799": "seggpt/modeling_seggpt.py:SegGptDecoderHead", "5800": "seggpt/modeling_seggpt.py:SegGptDecoder", "5801": "seggpt/modeling_seggpt.py:SegGptPreTrainedModel", "5802": "seggpt/modeling_seggpt.py:SegGptModel", "5803": "seggpt/modeling_seggpt.py:patchify", "5804": "seggpt/modeling_seggpt.py:unpatchify", "5805": "seggpt/modeling_seggpt.py:SegGptLoss", "5806": "seggpt/modeling_seggpt.py:SegGptForImageSegmentation", "5807": "dia/modeling_dia.py:DiaPreTrainedModel", "5808": "dia/modeling_dia.py:DiaMultiChannelEmbedding", "5809": "dia/modeling_dia.py:DiaMLP", "5810": "dia/modeling_dia.py:DiaRMSNorm", "5811": "dia/modeling_dia.py:DiaRotaryEmbedding", "5812": "dia/modeling_dia.py:rotate_half", "5813": "dia/modeling_dia.py:apply_rotary_pos_emb", "5814": "dia/modeling_dia.py:repeat_kv", "5815": "dia/modeling_dia.py:eager_attention_forward", "5816": "dia/modeling_dia.py:DiaSelfAttention", "5817": "dia/modeling_dia.py:DiaCrossAttention", "5818": "dia/modeling_dia.py:DiaEncoderLayer", "5819": "dia/modeling_dia.py:DiaEncoder", "5820": "dia/modeling_dia.py:DiaDecoderLayer", "5821": "dia/modeling_dia.py:DiaDecoder", "5822": "dia/modeling_dia.py:DiaModel", "5823": "dia/modeling_dia.py:DiaForConditionalGeneration", "5824": "pegasus_x/modeling_pegasus_x.py:DimensionInfo", "5825": "pegasus_x/modeling_pegasus_x.py:shift_tokens_right", "5826": "pegasus_x/modeling_pegasus_x.py:PegasusXScaledWordEmbedding", "5827": "pegasus_x/modeling_pegasus_x.py:PegasusXSinusoidalPositionalEmbedding", "5828": "pegasus_x/modeling_pegasus_x.py:eager_attention_forward", "5829": "pegasus_x/modeling_pegasus_x.py:PegasusXAttention", "5830": "pegasus_x/modeling_pegasus_x.py:PegasusXGlobalLocalAttention", "5831": "pegasus_x/modeling_pegasus_x.py:PegasusXEncoderLayer", "5832": "pegasus_x/modeling_pegasus_x.py:PegasusXDecoderLayer", "5833": "pegasus_x/modeling_pegasus_x.py:PegasusXPreTrainedModel", "5834": "pegasus_x/modeling_pegasus_x.py:PegasusXEncoder", "5835": "pegasus_x/modeling_pegasus_x.py:PegasusXDecoder", "5836": "pegasus_x/modeling_pegasus_x.py:PegasusXModel", "5837": "pegasus_x/modeling_pegasus_x.py:PegasusXForConditionalGeneration", "5838": "pegasus_x/modeling_pegasus_x.py:PegasusXDecoderWrapper", "5839": "speech_to_text/modeling_speech_to_text.py:shift_tokens_right", "5840": "speech_to_text/modeling_speech_to_text.py:Conv1dSubsampler", "5841": "speech_to_text/modeling_speech_to_text.py:Speech2TextSinusoidalPositionalEmbedding", "5842": "speech_to_text/modeling_speech_to_text.py:eager_attention_forward", "5843": "speech_to_text/modeling_speech_to_text.py:Speech2TextAttention", "5844": "speech_to_text/modeling_speech_to_text.py:Speech2TextEncoderLayer", "5845": "speech_to_text/modeling_speech_to_text.py:Speech2TextDecoderLayer", "5846": "speech_to_text/modeling_speech_to_text.py:Speech2TextPreTrainedModel", "5847": "speech_to_text/modeling_speech_to_text.py:Speech2TextEncoder", "5848": "speech_to_text/modeling_speech_to_text.py:Speech2TextDecoder", "5849": "speech_to_text/modeling_speech_to_text.py:Speech2TextModel", "5850": "speech_to_text/modeling_speech_to_text.py:Speech2TextForConditionalGeneration", "5851": "nemotron/modeling_nemotron.py:_cast_if_autocast_enabled", "5852": "nemotron/modeling_nemotron.py:NemotronLayerNorm1P", "5853": "nemotron/modeling_nemotron.py:NemotronRotaryEmbedding", "5854": "nemotron/modeling_nemotron.py:rotate_half", "5855": "nemotron/modeling_nemotron.py:apply_rotary_pos_emb", "5856": "nemotron/modeling_nemotron.py:NemotronMLP", "5857": "nemotron/modeling_nemotron.py:repeat_kv", "5858": "nemotron/modeling_nemotron.py:NemotronAttention", "5859": "nemotron/modeling_nemotron.py:NemotronFlashAttention2", "5860": "nemotron/modeling_nemotron.py:NemotronSdpaAttention", "5861": "nemotron/modeling_nemotron.py:NemotronDecoderLayer", "5862": "nemotron/modeling_nemotron.py:NemotronPreTrainedModel", "5863": "nemotron/modeling_nemotron.py:NemotronModel", "5864": "nemotron/modeling_nemotron.py:NemotronForCausalLM", "5865": "nemotron/modeling_nemotron.py:NemotronForSequenceClassification", "5866": "nemotron/modeling_nemotron.py:NemotronForQuestionAnswering", "5867": "nemotron/modeling_nemotron.py:NemotronForTokenClassification", "5868": "lilt/modeling_lilt.py:LiltTextEmbeddings", "5869": "lilt/modeling_lilt.py:LiltLayoutEmbeddings", "5870": "lilt/modeling_lilt.py:LiltSelfAttention", "5871": "lilt/modeling_lilt.py:LiltSelfOutput", "5872": "lilt/modeling_lilt.py:LiltAttention", "5873": "lilt/modeling_lilt.py:LiltIntermediate", "5874": "lilt/modeling_lilt.py:LiltOutput", "5875": "lilt/modeling_lilt.py:LiltLayer", "5876": "lilt/modeling_lilt.py:LiltEncoder", "5877": "lilt/modeling_lilt.py:LiltPooler", "5878": "lilt/modeling_lilt.py:LiltPreTrainedModel", "5879": "lilt/modeling_lilt.py:LiltModel", "5880": "lilt/modeling_lilt.py:LiltForSequenceClassification", "5881": "lilt/modeling_lilt.py:LiltForTokenClassification", "5882": "lilt/modeling_lilt.py:LiltClassificationHead", "5883": "lilt/modeling_lilt.py:LiltForQuestionAnswering", "5884": "zamba/modeling_zamba.py:ZambaRMSNorm", "5885": "zamba/modeling_zamba.py:repeat_kv", "5886": "zamba/modeling_zamba.py:ZambaHybridDynamicCache", "5887": "zamba/modeling_zamba.py:eager_attention_forward", "5888": "zamba/modeling_zamba.py:ZambaAttention", "5889": "zamba/modeling_zamba.py:ZambaMambaMixer", "5890": "zamba/modeling_zamba.py:ZambaMLP", "5891": "zamba/modeling_zamba.py:ZambaAttentionDecoderLayer", "5892": "zamba/modeling_zamba.py:ZambaMambaDecoderLayer", "5893": "zamba/modeling_zamba.py:ZambaHybridLayer", "5894": "zamba/modeling_zamba.py:ZambaPreTrainedModel", "5895": "zamba/modeling_zamba.py:ZambaModel", "5896": "zamba/modeling_zamba.py:ZambaForCausalLM", "5897": "zamba/modeling_zamba.py:ZambaForSequenceClassification", "5898": "whisper/modeling_whisper.py:sinusoids", "5899": "whisper/modeling_whisper.py:shift_tokens_right", "5900": "whisper/modeling_whisper.py:_compute_mask_indices", "5901": "whisper/modeling_whisper.py:WhisperPositionalEmbedding", "5902": "whisper/modeling_whisper.py:eager_attention_forward", "5903": "whisper/modeling_whisper.py:WhisperAttention", "5904": "whisper/modeling_whisper.py:WhisperEncoderLayer", "5905": "whisper/modeling_whisper.py:WhisperDecoderLayer", "5906": "whisper/modeling_whisper.py:WhisperPreTrainedModel", "5907": "whisper/modeling_whisper.py:WhisperEncoder", "5908": "whisper/modeling_whisper.py:WhisperDecoder", "5909": "whisper/modeling_whisper.py:WhisperModel", "5910": "whisper/modeling_whisper.py:WhisperForConditionalGeneration", "5911": "whisper/modeling_whisper.py:WhisperDecoderWrapper", "5912": "whisper/modeling_whisper.py:WhisperForCausalLM", "5913": "whisper/modeling_whisper.py:WhisperForAudioClassification", "5914": "granite_speech/modeling_granite_speech.py:GraniteSpeechCausalLMOutputWithPast", "5915": "granite_speech/modeling_granite_speech.py:GraniteSpeechEncoderProjector", "5916": "granite_speech/modeling_granite_speech.py:GraniteSpeechConformerFeedForward", "5917": "granite_speech/modeling_granite_speech.py:GraniteSpeechConformerAttention", "5918": "granite_speech/modeling_granite_speech.py:GraniteSpeechConformerDepthWiseConv1d", "5919": "granite_speech/modeling_granite_speech.py:GraniteSpeechConformerConvModule", "5920": "granite_speech/modeling_granite_speech.py:GraniteSpeechConformerBlock", "5921": "granite_speech/modeling_granite_speech.py:GraniteSpeechCTCEncoder", "5922": "granite_speech/modeling_granite_speech.py:GraniteSpeechPreTrainedModel", "5923": "granite_speech/modeling_granite_speech.py:GraniteSpeechForConditionalGeneration", "5924": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3RMSNorm", "5925": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3RotaryEmbedding", "5926": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3MLP", "5927": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3TopkRouter", "5928": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3MoE", "5929": "deepseek_v3/modeling_deepseek_v3.py:rotate_half", "5930": "deepseek_v3/modeling_deepseek_v3.py:apply_rotary_pos_emb", "5931": "deepseek_v3/modeling_deepseek_v3.py:repeat_kv", "5932": "deepseek_v3/modeling_deepseek_v3.py:eager_attention_forward", "5933": "deepseek_v3/modeling_deepseek_v3.py:apply_rotary_pos_emb_interleave", "5934": "deepseek_v3/modeling_deepseek_v3.py:yarn_get_mscale", "5935": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3Attention", "5936": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3DecoderLayer", "5937": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3PreTrainedModel", "5938": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3Model", "5939": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3ForCausalLM", "5940": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3ForSequenceClassification", "5941": "deepseek_v3/modeling_deepseek_v3.py:DeepseekV3ForTokenClassification", "5942": "rwkv/modeling_rwkv.py:load_wkv_cuda_kernel", "5943": "rwkv/modeling_rwkv.py:RwkvLinearAttention", "5944": "rwkv/modeling_rwkv.py:rwkv_linear_attention_cpu", "5945": "rwkv/modeling_rwkv.py:rwkv_linear_attention", "5946": "rwkv/modeling_rwkv.py:RwkvSelfAttention", "5947": "rwkv/modeling_rwkv.py:RwkvFeedForward", "5948": "rwkv/modeling_rwkv.py:RwkvBlock", "5949": "rwkv/modeling_rwkv.py:RwkvPreTrainedModel", "5950": "rwkv/modeling_rwkv.py:RwkvOutput", "5951": "rwkv/modeling_rwkv.py:RwkvCausalLMOutput", "5952": "rwkv/modeling_rwkv.py:RwkvModel", "5953": "rwkv/modeling_rwkv.py:RwkvForCausalLM", "5954": "bamba/modeling_bamba.py:BambaFlashAttentionKwargs", "5955": "bamba/modeling_bamba.py:HybridMambaAttentionDynamicCache", "5956": "bamba/modeling_bamba.py:BambaRotaryEmbedding", "5957": "bamba/modeling_bamba.py:rotate_half", "5958": "bamba/modeling_bamba.py:repeat_kv", "5959": "bamba/modeling_bamba.py:eager_attention_forward", "5960": "bamba/modeling_bamba.py:apply_rotary_pos_emb", "5961": "bamba/modeling_bamba.py:BambaAttention", "5962": "bamba/modeling_bamba.py:BambaRMSNormGated", "5963": "bamba/modeling_bamba.py:pad_tensor_by_size", "5964": "bamba/modeling_bamba.py:reshape_into_chunks", "5965": "bamba/modeling_bamba.py:segment_sum", "5966": "bamba/modeling_bamba.py:apply_mask_to_padding_states", "5967": "bamba/modeling_bamba.py:BambaMixer", "5968": "bamba/modeling_bamba.py:BambaMLP", "5969": "bamba/modeling_bamba.py:BambaRMSNorm", "5970": "bamba/modeling_bamba.py:BambaDecoderLayer", "5971": "bamba/modeling_bamba.py:BambaPreTrainedModel", "5972": "bamba/modeling_bamba.py:BambaModel", "5973": "bamba/modeling_bamba.py:BambaForCausalLM", "5974": "olmo2/modeling_olmo2.py:Olmo2RMSNorm", "5975": "olmo2/modeling_olmo2.py:repeat_kv", "5976": "olmo2/modeling_olmo2.py:eager_attention_forward", "5977": "olmo2/modeling_olmo2.py:apply_rotary_pos_emb", "5978": "olmo2/modeling_olmo2.py:rotate_half", "5979": "olmo2/modeling_olmo2.py:Olmo2Attention", "5980": "olmo2/modeling_olmo2.py:Olmo2MLP", "5981": "olmo2/modeling_olmo2.py:Olmo2DecoderLayer", "5982": "olmo2/modeling_olmo2.py:Olmo2RotaryEmbedding", "5983": "olmo2/modeling_olmo2.py:Olmo2PreTrainedModel", "5984": "olmo2/modeling_olmo2.py:Olmo2Model", "5985": "olmo2/modeling_olmo2.py:Olmo2ForCausalLM", "5986": "blip_2/modeling_blip_2.py:Blip2ForConditionalGenerationModelOutput", "5987": "blip_2/modeling_blip_2.py:Blip2ImageTextMatchingModelOutput", "5988": "blip_2/modeling_blip_2.py:Blip2TextModelOutput", "5989": "blip_2/modeling_blip_2.py:Blip2VisionModelOutput", "5990": "blip_2/modeling_blip_2.py:Blip2VisionEmbeddings", "5991": "blip_2/modeling_blip_2.py:eager_attention_forward", "5992": "blip_2/modeling_blip_2.py:Blip2Attention", "5993": "blip_2/modeling_blip_2.py:Blip2MLP", "5994": "blip_2/modeling_blip_2.py:Blip2EncoderLayer", "5995": "blip_2/modeling_blip_2.py:Blip2PreTrainedModel", "5996": "blip_2/modeling_blip_2.py:Blip2Encoder", "5997": "blip_2/modeling_blip_2.py:Blip2VisionModel", "5998": "blip_2/modeling_blip_2.py:Blip2QFormerMultiHeadAttention", "5999": "blip_2/modeling_blip_2.py:Blip2QFormerSelfOutput", "6000": "blip_2/modeling_blip_2.py:Blip2QFormerAttention", "6001": "blip_2/modeling_blip_2.py:Blip2QFormerIntermediate", "6002": "blip_2/modeling_blip_2.py:Blip2QFormerOutput", "6003": "blip_2/modeling_blip_2.py:Blip2QFormerLayer", "6004": "blip_2/modeling_blip_2.py:Blip2QFormerEncoder", "6005": "blip_2/modeling_blip_2.py:Blip2TextEmbeddings", "6006": "blip_2/modeling_blip_2.py:Blip2QFormerModel", "6007": "blip_2/modeling_blip_2.py:Blip2Model", "6008": "blip_2/modeling_blip_2.py:Blip2TextModelWithProjection", "6009": "blip_2/modeling_blip_2.py:Blip2VisionModelWithProjection", "6010": "blip_2/modeling_blip_2.py:Blip2ForConditionalGeneration", "6011": "blip_2/modeling_blip_2.py:Blip2ForImageTextRetrieval", "6012": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TGenerationOutput", "6013": "seamless_m4t/modeling_seamless_m4t.py:shift_tokens_right", "6014": "seamless_m4t/modeling_seamless_m4t.py:_compute_new_attention_mask", "6015": "seamless_m4t/modeling_seamless_m4t.py:format_speech_generation_kwargs", "6016": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerPositionalConvEmbedding", "6017": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerRotaryPositionalEmbedding", "6018": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerRelPositionalEmbedding", "6019": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerSamePadLayer", "6020": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerFeatureProjection", "6021": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerFeedForward", "6022": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerConvolutionModule", "6023": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerSelfAttention", "6024": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerEncoderLayer", "6025": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerEncoder", "6026": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerAdapterLayer", "6027": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TConformerAdapter", "6028": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TScaledWordEmbedding", "6029": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TSinusoidalPositionalEmbedding", "6030": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TAttention", "6031": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TFeedForwardNetwork", "6032": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TEncoderLayer", "6033": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TDecoderLayer", "6034": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TPreTrainedModel", "6035": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TSpeechEncoder", "6036": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TEncoder", "6037": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TDecoder", "6038": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TTextToUnitModel", "6039": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TTextToUnitForConditionalGeneration", "6040": "seamless_m4t/modeling_seamless_m4t.py:HifiGanResidualBlock", "6041": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TVariancePredictor", "6042": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4THifiGan", "6043": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TCodeHifiGan", "6044": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TForTextToText", "6045": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TForSpeechToText", "6046": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TForTextToSpeech", "6047": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TForSpeechToSpeech", "6048": "seamless_m4t/modeling_seamless_m4t.py:SeamlessM4TModel", "6049": "instructblip/modeling_instructblip.py:InstructBlipForConditionalGenerationModelOutput", "6050": "instructblip/modeling_instructblip.py:InstructBlipVisionEmbeddings", "6051": "instructblip/modeling_instructblip.py:eager_attention_forward", "6052": "instructblip/modeling_instructblip.py:InstructBlipAttention", "6053": "instructblip/modeling_instructblip.py:InstructBlipMLP", "6054": "instructblip/modeling_instructblip.py:InstructBlipEncoderLayer", "6055": "instructblip/modeling_instructblip.py:InstructBlipPreTrainedModel", "6056": "instructblip/modeling_instructblip.py:InstructBlipEncoder", "6057": "instructblip/modeling_instructblip.py:InstructBlipVisionModel", "6058": "instructblip/modeling_instructblip.py:InstructBlipQFormerMultiHeadAttention", "6059": "instructblip/modeling_instructblip.py:InstructBlipQFormerSelfOutput", "6060": "instructblip/modeling_instructblip.py:InstructBlipQFormerAttention", "6061": "instructblip/modeling_instructblip.py:InstructBlipQFormerIntermediate", "6062": "instructblip/modeling_instructblip.py:InstructBlipQFormerOutput", "6063": "instructblip/modeling_instructblip.py:InstructBlipQFormerLayer", "6064": "instructblip/modeling_instructblip.py:InstructBlipQFormerEncoder", "6065": "instructblip/modeling_instructblip.py:InstructBlipQFormerEmbeddings", "6066": "instructblip/modeling_instructblip.py:InstructBlipQFormerModel", "6067": "instructblip/modeling_instructblip.py:InstructBlipModel", "6068": "instructblip/modeling_instructblip.py:InstructBlipForConditionalGeneration", "6069": "vaultgemma/modeling_vaultgemma.py:VaultGemmaRMSNorm", "6070": "vaultgemma/modeling_vaultgemma.py:VaultGemmaMLP", "6071": "vaultgemma/modeling_vaultgemma.py:rotate_half", "6072": "vaultgemma/modeling_vaultgemma.py:apply_rotary_pos_emb", "6073": "vaultgemma/modeling_vaultgemma.py:repeat_kv", "6074": "vaultgemma/modeling_vaultgemma.py:eager_attention_forward", "6075": "vaultgemma/modeling_vaultgemma.py:VaultGemmaAttention", "6076": "vaultgemma/modeling_vaultgemma.py:VaultGemmaDecoderLayer", "6077": "vaultgemma/modeling_vaultgemma.py:VaultGemmaRotaryEmbedding", "6078": "vaultgemma/modeling_vaultgemma.py:VaultGemmaPreTrainedModel", "6079": "vaultgemma/modeling_vaultgemma.py:VaultGemmaModel", "6080": "vaultgemma/modeling_vaultgemma.py:VaultGemmaForCausalLM", "6081": "mpnet/modeling_mpnet.py:MPNetPreTrainedModel", "6082": "mpnet/modeling_mpnet.py:MPNetEmbeddings", "6083": "mpnet/modeling_mpnet.py:MPNetSelfAttention", "6084": "mpnet/modeling_mpnet.py:MPNetAttention", "6085": "mpnet/modeling_mpnet.py:MPNetIntermediate", "6086": "mpnet/modeling_mpnet.py:MPNetOutput", "6087": "mpnet/modeling_mpnet.py:MPNetLayer", "6088": "mpnet/modeling_mpnet.py:MPNetEncoder", "6089": "mpnet/modeling_mpnet.py:MPNetPooler", "6090": "mpnet/modeling_mpnet.py:MPNetModel", "6091": "mpnet/modeling_mpnet.py:MPNetForMaskedLM", "6092": "mpnet/modeling_mpnet.py:MPNetLMHead", "6093": "mpnet/modeling_mpnet.py:MPNetForSequenceClassification", "6094": "mpnet/modeling_mpnet.py:MPNetForMultipleChoice", "6095": "mpnet/modeling_mpnet.py:MPNetForTokenClassification", "6096": "mpnet/modeling_mpnet.py:MPNetClassificationHead", "6097": "mpnet/modeling_mpnet.py:MPNetForQuestionAnswering", "6098": "mpnet/modeling_mpnet.py:create_position_ids_from_input_ids", "6099": "jamba/modeling_jamba.py:load_balancing_loss_func", "6100": "jamba/modeling_jamba.py:JambaRMSNorm", "6101": "jamba/modeling_jamba.py:repeat_kv", "6102": "jamba/modeling_jamba.py:HybridMambaAttentionDynamicCache", "6103": "jamba/modeling_jamba.py:JambaAttention", "6104": "jamba/modeling_jamba.py:JambaFlashAttention2", "6105": "jamba/modeling_jamba.py:JambaSdpaAttention", "6106": "jamba/modeling_jamba.py:JambaMambaMixer", "6107": "jamba/modeling_jamba.py:JambaMLP", "6108": "jamba/modeling_jamba.py:JambaSparseMoeBlock", "6109": "jamba/modeling_jamba.py:JambaAttentionDecoderLayer", "6110": "jamba/modeling_jamba.py:JambaMambaDecoderLayer", "6111": "jamba/modeling_jamba.py:JambaPreTrainedModel", "6112": "jamba/modeling_jamba.py:JambaModel", "6113": "jamba/modeling_jamba.py:JambaForCausalLM", "6114": "jamba/modeling_jamba.py:JambaForSequenceClassification", "6115": "aimv2/modeling_aimv2.py:Aimv2Output", "6116": "aimv2/modeling_aimv2.py:Aimv2RMSNorm", "6117": "aimv2/modeling_aimv2.py:Aimv2MLP", "6118": "aimv2/modeling_aimv2.py:Aimv2VisionEmbeddings", "6119": "aimv2/modeling_aimv2.py:Aimv2TextEmbeddings", "6120": "aimv2/modeling_aimv2.py:eager_attention_forward", "6121": "aimv2/modeling_aimv2.py:Aimv2Attention", "6122": "aimv2/modeling_aimv2.py:Aimv2EncoderLayer", "6123": "aimv2/modeling_aimv2.py:Aimv2Encoder", "6124": "aimv2/modeling_aimv2.py:Aimv2AttentionPoolingHead", "6125": "aimv2/modeling_aimv2.py:Aimv2PreTrainedModel", "6126": "aimv2/modeling_aimv2.py:Aimv2VisionModel", "6127": "aimv2/modeling_aimv2.py:Aimv2TextModel", "6128": "aimv2/modeling_aimv2.py:_get_vector_norm", "6129": "aimv2/modeling_aimv2.py:Aimv2Model", "6130": "resnet/modeling_resnet.py:ResNetConvLayer", "6131": "resnet/modeling_resnet.py:ResNetEmbeddings", "6132": "resnet/modeling_resnet.py:ResNetShortCut", "6133": "resnet/modeling_resnet.py:ResNetBasicLayer", "6134": "resnet/modeling_resnet.py:ResNetBottleNeckLayer", "6135": "resnet/modeling_resnet.py:ResNetStage", "6136": "resnet/modeling_resnet.py:ResNetEncoder", "6137": "resnet/modeling_resnet.py:ResNetPreTrainedModel", "6138": "resnet/modeling_resnet.py:ResNetModel", "6139": "resnet/modeling_resnet.py:ResNetForImageClassification", "6140": "resnet/modeling_resnet.py:ResNetBackbone", "6141": "diffllama/modeling_diffllama.py:DiffLlamaMLP", "6142": "diffllama/modeling_diffllama.py:rotate_half", "6143": "diffllama/modeling_diffllama.py:apply_rotary_pos_emb", "6144": "diffllama/modeling_diffllama.py:repeat_kv", "6145": "diffllama/modeling_diffllama.py:lambda_init_fn", "6146": "diffllama/modeling_diffllama.py:DiffLlamaAttention", "6147": "diffllama/modeling_diffllama.py:DiffLlamaFlashAttention2", "6148": "diffllama/modeling_diffllama.py:DiffLlamaSdpaAttention", "6149": "diffllama/modeling_diffllama.py:DiffLlamaRMSNorm", "6150": "diffllama/modeling_diffllama.py:DiffLlamaDecoderLayer", "6151": "diffllama/modeling_diffllama.py:DiffLlamaPreTrainedModel", "6152": "diffllama/modeling_diffllama.py:DiffLlamaRotaryEmbedding", "6153": "diffllama/modeling_diffllama.py:DiffLlamaModel", "6154": "diffllama/modeling_diffllama.py:DiffLlamaForCausalLM", "6155": "diffllama/modeling_diffllama.py:DiffLlamaForSequenceClassification", "6156": "diffllama/modeling_diffllama.py:DiffLlamaForQuestionAnswering", "6157": "diffllama/modeling_diffllama.py:DiffLlamaForTokenClassification", "6158": "swinv2/modeling_swinv2.py:Swinv2EncoderOutput", "6159": "swinv2/modeling_swinv2.py:Swinv2ModelOutput", "6160": "swinv2/modeling_swinv2.py:Swinv2MaskedImageModelingOutput", "6161": "swinv2/modeling_swinv2.py:Swinv2ImageClassifierOutput", "6162": "swinv2/modeling_swinv2.py:window_partition", "6163": "swinv2/modeling_swinv2.py:window_reverse", "6164": "swinv2/modeling_swinv2.py:drop_path", "6165": "swinv2/modeling_swinv2.py:Swinv2DropPath", "6166": "swinv2/modeling_swinv2.py:Swinv2Embeddings", "6167": "swinv2/modeling_swinv2.py:Swinv2PatchEmbeddings", "6168": "swinv2/modeling_swinv2.py:Swinv2PatchMerging", "6169": "swinv2/modeling_swinv2.py:Swinv2SelfAttention", "6170": "swinv2/modeling_swinv2.py:Swinv2SelfOutput", "6171": "swinv2/modeling_swinv2.py:Swinv2Attention", "6172": "swinv2/modeling_swinv2.py:Swinv2Intermediate", "6173": "swinv2/modeling_swinv2.py:Swinv2Output", "6174": "swinv2/modeling_swinv2.py:Swinv2Layer", "6175": "swinv2/modeling_swinv2.py:Swinv2Stage", "6176": "swinv2/modeling_swinv2.py:Swinv2Encoder", "6177": "swinv2/modeling_swinv2.py:Swinv2PreTrainedModel", "6178": "swinv2/modeling_swinv2.py:Swinv2Model", "6179": "swinv2/modeling_swinv2.py:Swinv2ForMaskedImageModeling", "6180": "swinv2/modeling_swinv2.py:Swinv2ForImageClassification", "6181": "swinv2/modeling_swinv2.py:Swinv2Backbone", "6182": "rt_detr_v2/modeling_rt_detr_v2.py:multi_scale_deformable_attention_v2", "6183": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2MultiscaleDeformableAttention", "6184": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2MultiheadAttention", "6185": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2DecoderLayer", "6186": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2PreTrainedModel", "6187": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2DecoderOutput", "6188": "rt_detr_v2/modeling_rt_detr_v2.py:inverse_sigmoid", "6189": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2Decoder", "6190": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2ModelOutput", "6191": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2FrozenBatchNorm2d", "6192": "rt_detr_v2/modeling_rt_detr_v2.py:replace_batch_norm", "6193": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2ConvEncoder", "6194": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2ConvNormLayer", "6195": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2EncoderLayer", "6196": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2RepVggBlock", "6197": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2CSPRepLayer", "6198": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2Encoder", "6199": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2HybridEncoder", "6200": "rt_detr_v2/modeling_rt_detr_v2.py:get_contrastive_denoising_training_group", "6201": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2Model", "6202": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2MLPPredictionHead", "6203": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2ObjectDetectionOutput", "6204": "rt_detr_v2/modeling_rt_detr_v2.py:RTDetrV2ForObjectDetection", "6205": "ijepa/modeling_ijepa.py:IJepaPatchEmbeddings", "6206": "ijepa/modeling_ijepa.py:IJepaEmbeddings", "6207": "ijepa/modeling_ijepa.py:eager_attention_forward", "6208": "ijepa/modeling_ijepa.py:IJepaSelfAttention", "6209": "ijepa/modeling_ijepa.py:IJepaSelfOutput", "6210": "ijepa/modeling_ijepa.py:IJepaAttention", "6211": "ijepa/modeling_ijepa.py:IJepaIntermediate", "6212": "ijepa/modeling_ijepa.py:IJepaOutput", "6213": "ijepa/modeling_ijepa.py:IJepaLayer", "6214": "ijepa/modeling_ijepa.py:IJepaPreTrainedModel", "6215": "ijepa/modeling_ijepa.py:IJepaEncoder", "6216": "ijepa/modeling_ijepa.py:IJepaPooler", "6217": "ijepa/modeling_ijepa.py:IJepaModel", "6218": "ijepa/modeling_ijepa.py:IJepaForImageClassification", "6219": "mbart/modeling_mbart.py:shift_tokens_right", "6220": "mbart/modeling_mbart.py:MBartLearnedPositionalEmbedding", "6221": "mbart/modeling_mbart.py:MBartScaledWordEmbedding", "6222": "mbart/modeling_mbart.py:eager_attention_forward", "6223": "mbart/modeling_mbart.py:MBartAttention", "6224": "mbart/modeling_mbart.py:MBartEncoderLayer", "6225": "mbart/modeling_mbart.py:MBartDecoderLayer", "6226": "mbart/modeling_mbart.py:MBartClassificationHead", "6227": "mbart/modeling_mbart.py:MBartPreTrainedModel", "6228": "mbart/modeling_mbart.py:MBartEncoder", "6229": "mbart/modeling_mbart.py:MBartDecoder", "6230": "mbart/modeling_mbart.py:MBartModel", "6231": "mbart/modeling_mbart.py:MBartForConditionalGeneration", "6232": "mbart/modeling_mbart.py:MBartForSequenceClassification", "6233": "mbart/modeling_mbart.py:MBartForQuestionAnswering", "6234": "mbart/modeling_mbart.py:MBartDecoderWrapper", "6235": "mbart/modeling_mbart.py:MBartForCausalLM", "6236": "beit/modeling_beit.py:BeitModelOutputWithPooling", "6237": "beit/modeling_beit.py:drop_path", "6238": "beit/modeling_beit.py:BeitDropPath", "6239": "beit/modeling_beit.py:BeitEmbeddings", "6240": "beit/modeling_beit.py:BeitPatchEmbeddings", "6241": "beit/modeling_beit.py:BeitSelfAttention", "6242": "beit/modeling_beit.py:BeitSdpaSelfAttention", "6243": "beit/modeling_beit.py:BeitSelfOutput", "6244": "beit/modeling_beit.py:BeitAttention", "6245": "beit/modeling_beit.py:BeitIntermediate", "6246": "beit/modeling_beit.py:BeitOutput", "6247": "beit/modeling_beit.py:BeitLayer", "6248": "beit/modeling_beit.py:BeitRelativePositionBias", "6249": "beit/modeling_beit.py:BeitEncoder", "6250": "beit/modeling_beit.py:BeitPreTrainedModel", "6251": "beit/modeling_beit.py:BeitModel", "6252": "beit/modeling_beit.py:BeitPooler", "6253": "beit/modeling_beit.py:BeitForMaskedImageModeling", "6254": "beit/modeling_beit.py:BeitForImageClassification", "6255": "beit/modeling_beit.py:BeitConvModule", "6256": "beit/modeling_beit.py:BeitPyramidPoolingBlock", "6257": "beit/modeling_beit.py:BeitPyramidPoolingModule", "6258": "beit/modeling_beit.py:BeitUperHead", "6259": "beit/modeling_beit.py:BeitFCNHead", "6260": "beit/modeling_beit.py:BeitForSemanticSegmentation", "6261": "beit/modeling_beit.py:BeitBackbone", "6262": "align/modeling_align.py:AlignVisionModelOutput", "6263": "align/modeling_align.py:AlignTextModelOutput", "6264": "align/modeling_align.py:AlignOutput", "6265": "align/modeling_align.py:contrastive_loss", "6266": "align/modeling_align.py:align_loss", "6267": "align/modeling_align.py:round_filters", "6268": "align/modeling_align.py:correct_pad", "6269": "align/modeling_align.py:AlignVisionEmbeddings", "6270": "align/modeling_align.py:AlignVisionDepthwiseConv2d", "6271": "align/modeling_align.py:AlignVisionExpansionLayer", "6272": "align/modeling_align.py:AlignVisionDepthwiseLayer", "6273": "align/modeling_align.py:AlignVisionSqueezeExciteLayer", "6274": "align/modeling_align.py:AlignVisionFinalBlockLayer", "6275": "align/modeling_align.py:AlignVisionBlock", "6276": "align/modeling_align.py:AlignVisionEncoder", "6277": "align/modeling_align.py:AlignTextEmbeddings", "6278": "align/modeling_align.py:eager_attention_forward", "6279": "align/modeling_align.py:AlignTextSelfAttention", "6280": "align/modeling_align.py:AlignTextSelfOutput", "6281": "align/modeling_align.py:AlignTextAttention", "6282": "align/modeling_align.py:AlignTextIntermediate", "6283": "align/modeling_align.py:AlignTextOutput", "6284": "align/modeling_align.py:AlignTextLayer", "6285": "align/modeling_align.py:AlignTextEncoder", "6286": "align/modeling_align.py:AlignTextPooler", "6287": "align/modeling_align.py:AlignPreTrainedModel", "6288": "align/modeling_align.py:AlignTextModel", "6289": "align/modeling_align.py:AlignVisionModel", "6290": "align/modeling_align.py:AlignModel", "6291": "video_llava/modeling_video_llava.py:VideoLlavaModelOutputWithPast", "6292": "video_llava/modeling_video_llava.py:VideoLlavaCausalLMOutputWithPast", "6293": "video_llava/modeling_video_llava.py:VideoLlavaMultiModalProjector", "6294": "video_llava/modeling_video_llava.py:VideoLlavaPreTrainedModel", "6295": "video_llava/modeling_video_llava.py:VideoLlavaModel", "6296": "video_llava/modeling_video_llava.py:VideoLlavaForConditionalGeneration", "6297": "x_clip/modeling_x_clip.py:contrastive_loss", "6298": "x_clip/modeling_x_clip.py:x_clip_loss", "6299": "x_clip/modeling_x_clip.py:XCLIPOutput", "6300": "x_clip/modeling_x_clip.py:XCLIPVisionEmbeddings", "6301": "x_clip/modeling_x_clip.py:XCLIPTextEmbeddings", "6302": "x_clip/modeling_x_clip.py:eager_attention_forward", "6303": "x_clip/modeling_x_clip.py:XCLIPAttention", "6304": "x_clip/modeling_x_clip.py:XCLIPMLP", "6305": "x_clip/modeling_x_clip.py:XCLIPEncoderLayer", "6306": "x_clip/modeling_x_clip.py:drop_path", "6307": "x_clip/modeling_x_clip.py:XCLIPDropPath", "6308": "x_clip/modeling_x_clip.py:XCLIPVisionEncoderLayer", "6309": "x_clip/modeling_x_clip.py:XCLIPPreTrainedModel", "6310": "x_clip/modeling_x_clip.py:XCLIPEncoder", "6311": "x_clip/modeling_x_clip.py:XCLIPTextTransformer", "6312": "x_clip/modeling_x_clip.py:XCLIPTextModel", "6313": "x_clip/modeling_x_clip.py:XCLIPVisionEncoder", "6314": "x_clip/modeling_x_clip.py:XCLIPVisionTransformer", "6315": "x_clip/modeling_x_clip.py:XCLIPVisionModel", "6316": "x_clip/modeling_x_clip.py:XCLIPMultiframeIntegrationTransformer", "6317": "x_clip/modeling_x_clip.py:XCLIPCrossAttention", "6318": "x_clip/modeling_x_clip.py:PromptGeneratorLayer", "6319": "x_clip/modeling_x_clip.py:XCLIPPromptGenerator", "6320": "x_clip/modeling_x_clip.py:XCLIPModel", "6321": "levit/modeling_levit.py:LevitForImageClassificationWithTeacherOutput", "6322": "levit/modeling_levit.py:LevitConvEmbeddings", "6323": "levit/modeling_levit.py:LevitPatchEmbeddings", "6324": "levit/modeling_levit.py:MLPLayerWithBN", "6325": "levit/modeling_levit.py:LevitSubsample", "6326": "levit/modeling_levit.py:LevitAttention", "6327": "levit/modeling_levit.py:LevitAttentionSubsample", "6328": "levit/modeling_levit.py:LevitMLPLayer", "6329": "levit/modeling_levit.py:LevitResidualLayer", "6330": "levit/modeling_levit.py:LevitStage", "6331": "levit/modeling_levit.py:LevitEncoder", "6332": "levit/modeling_levit.py:LevitClassificationLayer", "6333": "levit/modeling_levit.py:LevitPreTrainedModel", "6334": "levit/modeling_levit.py:LevitModel", "6335": "levit/modeling_levit.py:LevitForImageClassification", "6336": "levit/modeling_levit.py:LevitForImageClassificationWithTeacher", "6337": "smollm3/modeling_smollm3.py:rotate_half", "6338": "smollm3/modeling_smollm3.py:apply_rotary_pos_emb", "6339": "smollm3/modeling_smollm3.py:repeat_kv", "6340": "smollm3/modeling_smollm3.py:eager_attention_forward", "6341": "smollm3/modeling_smollm3.py:SmolLM3Attention", "6342": "smollm3/modeling_smollm3.py:SmolLM3RMSNorm", "6343": "smollm3/modeling_smollm3.py:SmolLM3MLP", "6344": "smollm3/modeling_smollm3.py:SmolLM3DecoderLayer", "6345": "smollm3/modeling_smollm3.py:SmolLM3PreTrainedModel", "6346": "smollm3/modeling_smollm3.py:SmolLM3RotaryEmbedding", "6347": "smollm3/modeling_smollm3.py:SmolLM3Model", "6348": "smollm3/modeling_smollm3.py:SmolLM3ForCausalLM", "6349": "smollm3/modeling_smollm3.py:SmolLM3ForSequenceClassification", "6350": "smollm3/modeling_smollm3.py:SmolLM3ForTokenClassification", "6351": "smollm3/modeling_smollm3.py:SmolLM3ForQuestionAnswering", "6352": "clipseg/modeling_clipseg.py:contrastive_loss", "6353": "clipseg/modeling_clipseg.py:clipseg_loss", "6354": "clipseg/modeling_clipseg.py:CLIPSegOutput", "6355": "clipseg/modeling_clipseg.py:CLIPSegDecoderOutput", "6356": "clipseg/modeling_clipseg.py:CLIPSegImageSegmentationOutput", "6357": "clipseg/modeling_clipseg.py:CLIPSegVisionEmbeddings", "6358": "clipseg/modeling_clipseg.py:CLIPSegTextEmbeddings", "6359": "clipseg/modeling_clipseg.py:eager_attention_forward", "6360": "clipseg/modeling_clipseg.py:CLIPSegAttention", "6361": "clipseg/modeling_clipseg.py:CLIPSegMLP", "6362": "clipseg/modeling_clipseg.py:CLIPSegEncoderLayer", "6363": "clipseg/modeling_clipseg.py:CLIPSegPreTrainedModel", "6364": "clipseg/modeling_clipseg.py:CLIPSegEncoder", "6365": "clipseg/modeling_clipseg.py:CLIPSegTextTransformer", "6366": "clipseg/modeling_clipseg.py:CLIPSegTextModel", "6367": "clipseg/modeling_clipseg.py:CLIPSegVisionTransformer", "6368": "clipseg/modeling_clipseg.py:CLIPSegVisionModel", "6369": "clipseg/modeling_clipseg.py:CLIPSegModel", "6370": "clipseg/modeling_clipseg.py:CLIPSegDecoderLayer", "6371": "clipseg/modeling_clipseg.py:CLIPSegDecoder", "6372": "clipseg/modeling_clipseg.py:CLIPSegForImageSegmentation", "6373": "cohere2/modeling_cohere2.py:Cohere2RotaryEmbedding", "6374": "cohere2/modeling_cohere2.py:Cohere2LayerNorm", "6375": "cohere2/modeling_cohere2.py:repeat_kv", "6376": "cohere2/modeling_cohere2.py:eager_attention_forward", "6377": "cohere2/modeling_cohere2.py:rotate_half", "6378": "cohere2/modeling_cohere2.py:apply_rotary_pos_emb", "6379": "cohere2/modeling_cohere2.py:Cohere2Attention", "6380": "cohere2/modeling_cohere2.py:Cohere2MLP", "6381": "cohere2/modeling_cohere2.py:Cohere2DecoderLayer", "6382": "cohere2/modeling_cohere2.py:Cohere2PreTrainedModel", "6383": "cohere2/modeling_cohere2.py:Cohere2Model", "6384": "cohere2/modeling_cohere2.py:Cohere2ForCausalLM", "6385": "llava_next/modeling_llava_next.py:get_anyres_image_grid_shape", "6386": "llava_next/modeling_llava_next.py:image_size_to_num_patches", "6387": "llava_next/modeling_llava_next.py:unpad_image", "6388": "llava_next/modeling_llava_next.py:LlavaNextModelOutputWithPast", "6389": "llava_next/modeling_llava_next.py:LlavaNextCausalLMOutputWithPast", "6390": "llava_next/modeling_llava_next.py:LlavaNextMultiModalProjector", "6391": "llava_next/modeling_llava_next.py:LlavaNextPreTrainedModel", "6392": "llava_next/modeling_llava_next.py:LlavaNextModel", "6393": "llava_next/modeling_llava_next.py:LlavaNextForConditionalGeneration", "6394": "cpmant/modeling_cpmant.py:CpmAntLayerNorm", "6395": "cpmant/modeling_cpmant.py:CpmAntAttention", "6396": "cpmant/modeling_cpmant.py:CpmAntSelfAttentionBlock", "6397": "cpmant/modeling_cpmant.py:CpmAntDenseGatedACT", "6398": "cpmant/modeling_cpmant.py:CpmAntFeedForward", "6399": "cpmant/modeling_cpmant.py:CpmAntFFNBlock", "6400": "cpmant/modeling_cpmant.py:CpmAntTransformerBlock", "6401": "cpmant/modeling_cpmant.py:CpmAntEncoder", "6402": "cpmant/modeling_cpmant.py:CpmAntIntermediate", "6403": "cpmant/modeling_cpmant.py:CpmAntSegmentPositionEmbedding", "6404": "cpmant/modeling_cpmant.py:CpmAntOutput", "6405": "cpmant/modeling_cpmant.py:CpmAntPreTrainedModel", "6406": "cpmant/modeling_cpmant.py:CpmAntModel", "6407": "cpmant/modeling_cpmant.py:CpmAntForCausalLM", "6408": "sew_d/modeling_sew_d.py:_compute_mask_indices", "6409": "sew_d/modeling_sew_d.py:make_log_bucket_position", "6410": "sew_d/modeling_sew_d.py:build_relative_position", "6411": "sew_d/modeling_sew_d.py:c2p_dynamic_expand", "6412": "sew_d/modeling_sew_d.py:p2c_dynamic_expand", "6413": "sew_d/modeling_sew_d.py:pos_dynamic_expand", "6414": "sew_d/modeling_sew_d.py:get_mask", "6415": "sew_d/modeling_sew_d.py:SEWDNoLayerNormConvLayer", "6416": "sew_d/modeling_sew_d.py:SEWDLayerNormConvLayer", "6417": "sew_d/modeling_sew_d.py:SEWDGroupNormConvLayer", "6418": "sew_d/modeling_sew_d.py:SEWDPositionalConvEmbedding", "6419": "sew_d/modeling_sew_d.py:SEWDSamePadLayer", "6420": "sew_d/modeling_sew_d.py:SEWDUpsampling", "6421": "sew_d/modeling_sew_d.py:SEWDFeatureEncoder", "6422": "sew_d/modeling_sew_d.py:SEWDFeatureExtractor", "6423": "sew_d/modeling_sew_d.py:ContextPooler", "6424": "sew_d/modeling_sew_d.py:XSoftmax", "6425": "sew_d/modeling_sew_d.py:DropoutContext", "6426": "sew_d/modeling_sew_d.py:XDropout", "6427": "sew_d/modeling_sew_d.py:StableDropout", "6428": "sew_d/modeling_sew_d.py:SEWDSelfOutput", "6429": "sew_d/modeling_sew_d.py:DisentangledSelfAttention", "6430": "sew_d/modeling_sew_d.py:SEWDAttention", "6431": "sew_d/modeling_sew_d.py:SEWDIntermediate", "6432": "sew_d/modeling_sew_d.py:SEWDOutput", "6433": "sew_d/modeling_sew_d.py:SEWDLayer", "6434": "sew_d/modeling_sew_d.py:ConvLayer", "6435": "sew_d/modeling_sew_d.py:SEWDTransformerEncoder", "6436": "sew_d/modeling_sew_d.py:SEWDEncoder", "6437": "sew_d/modeling_sew_d.py:SEWDPreTrainedModel", "6438": "sew_d/modeling_sew_d.py:SEWDModel", "6439": "sew_d/modeling_sew_d.py:SEWDForCTC", "6440": "sew_d/modeling_sew_d.py:SEWDForSequenceClassification", "6441": "vivit/modeling_vivit.py:VivitTubeletEmbeddings", "6442": "vivit/modeling_vivit.py:VivitEmbeddings", "6443": "vivit/modeling_vivit.py:eager_attention_forward", "6444": "vivit/modeling_vivit.py:VivitSelfAttention", "6445": "vivit/modeling_vivit.py:VivitSelfOutput", "6446": "vivit/modeling_vivit.py:VivitAttention", "6447": "vivit/modeling_vivit.py:VivitIntermediate", "6448": "vivit/modeling_vivit.py:VivitOutput", "6449": "vivit/modeling_vivit.py:VivitLayer", "6450": "vivit/modeling_vivit.py:VivitEncoder", "6451": "vivit/modeling_vivit.py:VivitPooler", "6452": "vivit/modeling_vivit.py:VivitPreTrainedModel", "6453": "vivit/modeling_vivit.py:VivitModel", "6454": "vivit/modeling_vivit.py:VivitForVideoClassification", "6455": "biogpt/modeling_biogpt.py:BioGptLearnedPositionalEmbedding", "6456": "biogpt/modeling_biogpt.py:BioGptScaledWordEmbedding", "6457": "biogpt/modeling_biogpt.py:eager_attention_forward", "6458": "biogpt/modeling_biogpt.py:BioGptAttention", "6459": "biogpt/modeling_biogpt.py:BioGptDecoderLayer", "6460": "biogpt/modeling_biogpt.py:BioGptPreTrainedModel", "6461": "biogpt/modeling_biogpt.py:BioGptModel", "6462": "biogpt/modeling_biogpt.py:BioGptForCausalLM", "6463": "biogpt/modeling_biogpt.py:BioGptForTokenClassification", "6464": "biogpt/modeling_biogpt.py:BioGptForSequenceClassification", "6465": "yolos/modeling_yolos.py:YolosObjectDetectionOutput", "6466": "yolos/modeling_yolos.py:YolosEmbeddings", "6467": "yolos/modeling_yolos.py:InterpolateInitialPositionEmbeddings", "6468": "yolos/modeling_yolos.py:InterpolateMidPositionEmbeddings", "6469": "yolos/modeling_yolos.py:YolosPatchEmbeddings", "6470": "yolos/modeling_yolos.py:eager_attention_forward", "6471": "yolos/modeling_yolos.py:YolosSelfAttention", "6472": "yolos/modeling_yolos.py:YolosSelfOutput", "6473": "yolos/modeling_yolos.py:YolosAttention", "6474": "yolos/modeling_yolos.py:YolosIntermediate", "6475": "yolos/modeling_yolos.py:YolosOutput", "6476": "yolos/modeling_yolos.py:YolosLayer", "6477": "yolos/modeling_yolos.py:YolosEncoder", "6478": "yolos/modeling_yolos.py:YolosPreTrainedModel", "6479": "yolos/modeling_yolos.py:YolosModel", "6480": "yolos/modeling_yolos.py:YolosPooler", "6481": "yolos/modeling_yolos.py:YolosMLPPredictionHead", "6482": "yolos/modeling_yolos.py:YolosForObjectDetection", "6483": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatForPreTrainingOutput", "6484": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatSamePadLayer", "6485": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatPositionalConvEmbedding", "6486": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatNoLayerNormConvLayer", "6487": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatLayerNormConvLayer", "6488": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatGroupNormConvLayer", "6489": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatFeatureEncoder", "6490": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatFeatureProjection", "6491": "unispeech_sat/modeling_unispeech_sat.py:eager_attention_forward", "6492": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatAttention", "6493": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatFeedForward", "6494": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatEncoderLayer", "6495": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatEncoder", "6496": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatAttnAdapterLayer", "6497": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatEncoderLayerStableLayerNorm", "6498": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatEncoderStableLayerNorm", "6499": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatGumbelVectorQuantizer", "6500": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatPreTrainedModel", "6501": "unispeech_sat/modeling_unispeech_sat.py:_compute_mask_indices", "6502": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatModel", "6503": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatForPreTraining", "6504": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatForCTC", "6505": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatForSequenceClassification", "6506": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatForAudioFrameClassification", "6507": "unispeech_sat/modeling_unispeech_sat.py:AMSoftmaxLoss", "6508": "unispeech_sat/modeling_unispeech_sat.py:TDNNLayer", "6509": "unispeech_sat/modeling_unispeech_sat.py:UniSpeechSatForXVector", "6510": "patchtst/modeling_patchtst.py:eager_attention_forward", "6511": "patchtst/modeling_patchtst.py:PatchTSTAttention", "6512": "patchtst/modeling_patchtst.py:PatchTSTBatchNorm", "6513": "patchtst/modeling_patchtst.py:random_masking", "6514": "patchtst/modeling_patchtst.py:forecast_masking", "6515": "patchtst/modeling_patchtst.py:PatchTSTPatchify", "6516": "patchtst/modeling_patchtst.py:PatchTSTMasking", "6517": "patchtst/modeling_patchtst.py:PatchTSTEncoderLayer", "6518": "patchtst/modeling_patchtst.py:PatchTSTPreTrainedModel", "6519": "patchtst/modeling_patchtst.py:PatchTSTEmbedding", "6520": "patchtst/modeling_patchtst.py:PatchTSTPositionalEncoding", "6521": "patchtst/modeling_patchtst.py:PatchTSTEncoder", "6522": "patchtst/modeling_patchtst.py:PatchTSTModelOutput", "6523": "patchtst/modeling_patchtst.py:PatchTSTForPretrainingOutput", "6524": "patchtst/modeling_patchtst.py:PatchTSTForRegressionOutput", "6525": "patchtst/modeling_patchtst.py:PatchTSTForPredictionOutput", "6526": "patchtst/modeling_patchtst.py:PatchTSTForClassificationOutput", "6527": "patchtst/modeling_patchtst.py:SamplePatchTSTOutput", "6528": "patchtst/modeling_patchtst.py:nll", "6529": "patchtst/modeling_patchtst.py:weighted_average", "6530": "patchtst/modeling_patchtst.py:PatchTSTStdScaler", "6531": "patchtst/modeling_patchtst.py:PatchTSTMeanScaler", "6532": "patchtst/modeling_patchtst.py:PatchTSTNOPScaler", "6533": "patchtst/modeling_patchtst.py:PatchTSTScaler", "6534": "patchtst/modeling_patchtst.py:PatchTSTModel", "6535": "patchtst/modeling_patchtst.py:PatchTSTMaskPretrainHead", "6536": "patchtst/modeling_patchtst.py:PatchTSTForPretraining", "6537": "patchtst/modeling_patchtst.py:PatchTSTClassificationHead", "6538": "patchtst/modeling_patchtst.py:PatchTSTForClassification", "6539": "patchtst/modeling_patchtst.py:PatchTSTPredictionHead", "6540": "patchtst/modeling_patchtst.py:PatchTSTForPrediction", "6541": "patchtst/modeling_patchtst.py:PatchTSTRegressionHead", "6542": "patchtst/modeling_patchtst.py:PatchTSTForRegression", "6543": "siglip/modeling_siglip.py:_trunc_normal_", "6544": "siglip/modeling_siglip.py:trunc_normal_tf_", "6545": "siglip/modeling_siglip.py:variance_scaling_", "6546": "siglip/modeling_siglip.py:lecun_normal_", "6547": "siglip/modeling_siglip.py:default_flax_embed_init", "6548": "siglip/modeling_siglip.py:SiglipVisionModelOutput", "6549": "siglip/modeling_siglip.py:SiglipTextModelOutput", "6550": "siglip/modeling_siglip.py:SiglipOutput", "6551": "siglip/modeling_siglip.py:SiglipVisionEmbeddings", "6552": "siglip/modeling_siglip.py:SiglipTextEmbeddings", "6553": "siglip/modeling_siglip.py:eager_attention_forward", "6554": "siglip/modeling_siglip.py:SiglipAttention", "6555": "siglip/modeling_siglip.py:SiglipMLP", "6556": "siglip/modeling_siglip.py:SiglipEncoderLayer", "6557": "siglip/modeling_siglip.py:SiglipPreTrainedModel", "6558": "siglip/modeling_siglip.py:SiglipEncoder", "6559": "siglip/modeling_siglip.py:SiglipTextTransformer", "6560": "siglip/modeling_siglip.py:SiglipTextModel", "6561": "siglip/modeling_siglip.py:SiglipVisionTransformer", "6562": "siglip/modeling_siglip.py:SiglipMultiheadAttentionPoolingHead", "6563": "siglip/modeling_siglip.py:SiglipVisionModel", "6564": "siglip/modeling_siglip.py:SiglipModel", "6565": "siglip/modeling_siglip.py:SiglipForImageClassification", "6566": "qwen2/modeling_qwen2.py:Qwen2MLP", "6567": "qwen2/modeling_qwen2.py:rotate_half", "6568": "qwen2/modeling_qwen2.py:apply_rotary_pos_emb", "6569": "qwen2/modeling_qwen2.py:repeat_kv", "6570": "qwen2/modeling_qwen2.py:eager_attention_forward", "6571": "qwen2/modeling_qwen2.py:Qwen2Attention", "6572": "qwen2/modeling_qwen2.py:Qwen2RMSNorm", "6573": "qwen2/modeling_qwen2.py:Qwen2DecoderLayer", "6574": "qwen2/modeling_qwen2.py:Qwen2PreTrainedModel", "6575": "qwen2/modeling_qwen2.py:Qwen2RotaryEmbedding", "6576": "qwen2/modeling_qwen2.py:Qwen2Model", "6577": "qwen2/modeling_qwen2.py:Qwen2ForCausalLM", "6578": "qwen2/modeling_qwen2.py:Qwen2ForSequenceClassification", "6579": "qwen2/modeling_qwen2.py:Qwen2ForTokenClassification", "6580": "qwen2/modeling_qwen2.py:Qwen2ForQuestionAnswering", "6581": "cohere/modeling_cohere.py:CohereLayerNorm", "6582": "cohere/modeling_cohere.py:CohereRotaryEmbedding", "6583": "cohere/modeling_cohere.py:CohereMLP", "6584": "cohere/modeling_cohere.py:repeat_kv", "6585": "cohere/modeling_cohere.py:eager_attention_forward", "6586": "cohere/modeling_cohere.py:rotate_half", "6587": "cohere/modeling_cohere.py:apply_rotary_pos_emb", "6588": "cohere/modeling_cohere.py:CohereAttention", "6589": "cohere/modeling_cohere.py:CohereDecoderLayer", "6590": "cohere/modeling_cohere.py:CoherePreTrainedModel", "6591": "cohere/modeling_cohere.py:CohereModel", "6592": "cohere/modeling_cohere.py:CohereForCausalLM", "6593": "timm_wrapper/modeling_timm_wrapper.py:TimmWrapperModelOutput", "6594": "timm_wrapper/modeling_timm_wrapper.py:_create_timm_model_with_error_handling", "6595": "timm_wrapper/modeling_timm_wrapper.py:TimmWrapperPreTrainedModel", "6596": "timm_wrapper/modeling_timm_wrapper.py:TimmWrapperModel", "6597": "timm_wrapper/modeling_timm_wrapper.py:TimmWrapperForImageClassification", "6598": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniPreTrainedModel", "6599": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniPreTrainedModelForConditionalGeneration", "6600": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniThinkerCausalLMOutputWithPast", "6601": "qwen2_5_omni/modeling_qwen2_5_omni.py:repeat_kv", "6602": "qwen2_5_omni/modeling_qwen2_5_omni.py:eager_attention_forward", "6603": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniAudioAttention", "6604": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniAudioEncoderLayer", "6605": "qwen2_5_omni/modeling_qwen2_5_omni.py:SinusoidsPositionEmbedding", "6606": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniAudioEncoder", "6607": "qwen2_5_omni/modeling_qwen2_5_omni.py:rotate_half", "6608": "qwen2_5_omni/modeling_qwen2_5_omni.py:apply_rotary_pos_emb_vision", "6609": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniVisionAttention", "6610": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniMLP", "6611": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniVisionBlock", "6612": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5_VisionPatchEmbed", "6613": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5_VisionRotaryEmbedding", "6614": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniPatchMerger", "6615": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniVisionEncoder", "6616": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniRotaryEmbedding", "6617": "qwen2_5_omni/modeling_qwen2_5_omni.py:apply_multimodal_rotary_pos_emb", "6618": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniAttention", "6619": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2MLP", "6620": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniDecoderLayer", "6621": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniThinkerTextModel", "6622": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniThinkerForConditionalGeneration", "6623": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniTalkerCausalLMOutputWithPast", "6624": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniTalkerModel", "6625": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniTalkerForConditionalGeneration", "6626": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniDiTRotaryEmbedding", "6627": "qwen2_5_omni/modeling_qwen2_5_omni.py:TimeDelayNetBlock", "6628": "qwen2_5_omni/modeling_qwen2_5_omni.py:Res2NetBlock", "6629": "qwen2_5_omni/modeling_qwen2_5_omni.py:SqueezeExcitationBlock", "6630": "qwen2_5_omni/modeling_qwen2_5_omni.py:AttentiveStatisticsPooling", "6631": "qwen2_5_omni/modeling_qwen2_5_omni.py:SqueezeExcitationRes2NetBlock", "6632": "qwen2_5_omni/modeling_qwen2_5_omni.py:ECAPA_TimeDelayNet", "6633": "qwen2_5_omni/modeling_qwen2_5_omni.py:DiTInputEmbedding", "6634": "qwen2_5_omni/modeling_qwen2_5_omni.py:DiTCodecEmbedding", "6635": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5_OmniAdaLayerNormZero", "6636": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5_OmniAdaLayerNormZero_Final", "6637": "qwen2_5_omni/modeling_qwen2_5_omni.py:DiTMLP", "6638": "qwen2_5_omni/modeling_qwen2_5_omni.py:apply_rotary_pos_emb", "6639": "qwen2_5_omni/modeling_qwen2_5_omni.py:DiTAttention", "6640": "qwen2_5_omni/modeling_qwen2_5_omni.py:SinusPositionEmbedding", "6641": "qwen2_5_omni/modeling_qwen2_5_omni.py:DiTTimestepEmbedding", "6642": "qwen2_5_omni/modeling_qwen2_5_omni.py:DiTDecoderLayer", "6643": "qwen2_5_omni/modeling_qwen2_5_omni.py:SnakeBeta", "6644": "qwen2_5_omni/modeling_qwen2_5_omni.py:kaiser_sinc_filter1d", "6645": "qwen2_5_omni/modeling_qwen2_5_omni.py:UpSample1d", "6646": "qwen2_5_omni/modeling_qwen2_5_omni.py:DownSample1d", "6647": "qwen2_5_omni/modeling_qwen2_5_omni.py:TorchActivation1d", "6648": "qwen2_5_omni/modeling_qwen2_5_omni.py:AMPBlock", "6649": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniToken2WavBigVGANModel", "6650": "qwen2_5_omni/modeling_qwen2_5_omni.py:RungeKutta4ODESolver", "6651": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniToken2WavDiTModel", "6652": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniToken2WavModel", "6653": "qwen2_5_omni/modeling_qwen2_5_omni.py:Qwen2_5OmniForConditionalGeneration", "6654": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersPatchEmbeddings", "6655": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersEmbeddings", "6656": "dinov2_with_registers/modeling_dinov2_with_registers.py:eager_attention_forward", "6657": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersSelfAttention", "6658": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersSelfOutput", "6659": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersAttention", "6660": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersLayerScale", "6661": "dinov2_with_registers/modeling_dinov2_with_registers.py:drop_path", "6662": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersDropPath", "6663": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersMLP", "6664": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersSwiGLUFFN", "6665": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersLayer", "6666": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersEncoder", "6667": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersPreTrainedModel", "6668": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersModel", "6669": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersForImageClassification", "6670": "dinov2_with_registers/modeling_dinov2_with_registers.py:Dinov2WithRegistersBackbone", "6671": "deprecated/realm/modeling_realm.py:RealmEmbeddings", "6672": "deprecated/realm/modeling_realm.py:RealmSelfAttention", "6673": "deprecated/realm/modeling_realm.py:RealmSelfOutput", "6674": "deprecated/realm/modeling_realm.py:RealmAttention", "6675": "deprecated/realm/modeling_realm.py:RealmIntermediate", "6676": "deprecated/realm/modeling_realm.py:RealmOutput", "6677": "deprecated/realm/modeling_realm.py:RealmLayer", "6678": "deprecated/realm/modeling_realm.py:RealmEncoder", "6679": "deprecated/realm/modeling_realm.py:RealmPooler", "6680": "deprecated/realm/modeling_realm.py:RealmEmbedderOutput", "6681": "deprecated/realm/modeling_realm.py:RealmScorerOutput", "6682": "deprecated/realm/modeling_realm.py:RealmReaderOutput", "6683": "deprecated/realm/modeling_realm.py:RealmForOpenQAOutput", "6684": "deprecated/realm/modeling_realm.py:RealmPredictionHeadTransform", "6685": "deprecated/realm/modeling_realm.py:RealmLMPredictionHead", "6686": "deprecated/realm/modeling_realm.py:RealmOnlyMLMHead", "6687": "deprecated/realm/modeling_realm.py:RealmScorerProjection", "6688": "deprecated/realm/modeling_realm.py:RealmReaderProjection", "6689": "deprecated/realm/modeling_realm.py:RealmPreTrainedModel", "6690": "deprecated/realm/modeling_realm.py:RealmBertModel", "6691": "deprecated/realm/modeling_realm.py:RealmEmbedder", "6692": "deprecated/realm/modeling_realm.py:RealmScorer", "6693": "deprecated/realm/modeling_realm.py:RealmKnowledgeAugEncoder", "6694": "deprecated/realm/modeling_realm.py:RealmReader", "6695": "deprecated/realm/modeling_realm.py:RealmForOpenQA", "6696": "deprecated/transfo_xl/modeling_transfo_xl_utilities.py:ProjectedAdaptiveLogSoftmax", "6697": "deprecated/transfo_xl/modeling_transfo_xl.py:PositionalEmbedding", "6698": "deprecated/transfo_xl/modeling_transfo_xl.py:PositionwiseFF", "6699": "deprecated/transfo_xl/modeling_transfo_xl.py:RelPartialLearnableMultiHeadAttn", "6700": "deprecated/transfo_xl/modeling_transfo_xl.py:RelPartialLearnableDecoderLayer", "6701": "deprecated/transfo_xl/modeling_transfo_xl.py:AdaptiveEmbedding", "6702": "deprecated/transfo_xl/modeling_transfo_xl.py:TransfoXLPreTrainedModel", "6703": "deprecated/transfo_xl/modeling_transfo_xl.py:TransfoXLModelOutput", "6704": "deprecated/transfo_xl/modeling_transfo_xl.py:TransfoXLSequenceClassifierOutputWithPast", "6705": "deprecated/transfo_xl/modeling_transfo_xl.py:TransfoXLLMHeadModelOutput", "6706": "deprecated/transfo_xl/modeling_transfo_xl.py:TransfoXLModel", "6707": "deprecated/transfo_xl/modeling_transfo_xl.py:TransfoXLLMHeadModel", "6708": "deprecated/transfo_xl/modeling_transfo_xl.py:TransfoXLForSequenceClassification", "6709": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertEmbeddings", "6710": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertSelfAttention", "6711": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertSelfOutput", "6712": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertAttention", "6713": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertIntermediate", "6714": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertOutput", "6715": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertLayer", "6716": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertEncoder", "6717": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertPooler", "6718": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertPredictionHeadTransform", "6719": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertLMPredictionHead", "6720": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertOnlyMLMHead", "6721": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertOnlyNSPHead", "6722": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertPreTrainingHeads", "6723": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertPreTrainedModel", "6724": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertModel", "6725": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertLMHeadModel", "6726": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertForMaskedLM", "6727": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertForNextSentencePrediction", "6728": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertForSequenceClassification", "6729": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertForMultipleChoice", "6730": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertForTokenClassification", "6731": "deprecated/qdqbert/modeling_qdqbert.py:QDQBertForQuestionAnswering", "6732": "deprecated/tvlt/modeling_tvlt.py:TvltModelOutput", "6733": "deprecated/tvlt/modeling_tvlt.py:TvltDecoderOutput", "6734": "deprecated/tvlt/modeling_tvlt.py:TvltForPreTrainingOutput", "6735": "deprecated/tvlt/modeling_tvlt.py:generate_pixel_mask_noise", "6736": "deprecated/tvlt/modeling_tvlt.py:generate_audio_mask_noise", "6737": "deprecated/tvlt/modeling_tvlt.py:random_masking", "6738": "deprecated/tvlt/modeling_tvlt.py:TvltPixelEmbeddings", "6739": "deprecated/tvlt/modeling_tvlt.py:TvltAudioEmbeddings", "6740": "deprecated/tvlt/modeling_tvlt.py:TvltPixelPatchEmbeddings", "6741": "deprecated/tvlt/modeling_tvlt.py:TvltAudioPatchEmbeddings", "6742": "deprecated/tvlt/modeling_tvlt.py:TvltSelfAttention", "6743": "deprecated/tvlt/modeling_tvlt.py:TvltSelfOutput", "6744": "deprecated/tvlt/modeling_tvlt.py:TvltAttention", "6745": "deprecated/tvlt/modeling_tvlt.py:TvltIntermediate", "6746": "deprecated/tvlt/modeling_tvlt.py:TvltOutput", "6747": "deprecated/tvlt/modeling_tvlt.py:TvltLayer", "6748": "deprecated/tvlt/modeling_tvlt.py:TvltEncoder", "6749": "deprecated/tvlt/modeling_tvlt.py:TvltPreTrainedModel", "6750": "deprecated/tvlt/modeling_tvlt.py:TvltModel", "6751": "deprecated/tvlt/modeling_tvlt.py:TvltDecoder", "6752": "deprecated/tvlt/modeling_tvlt.py:TvltForPreTraining", "6753": "deprecated/tvlt/modeling_tvlt.py:TvltPooler", "6754": "deprecated/tvlt/modeling_tvlt.py:TvltMatchingHead", "6755": "deprecated/tvlt/modeling_tvlt.py:TvltMAEHead", "6756": "deprecated/tvlt/modeling_tvlt.py:TvltForAudioVisualClassification", "6757": "deprecated/deta/modeling_deta.py:load_cuda_kernels", "6758": "deprecated/deta/modeling_deta.py:MultiScaleDeformableAttentionFunction", "6759": "deprecated/deta/modeling_deta.py:DetaDecoderOutput", "6760": "deprecated/deta/modeling_deta.py:DetaModelOutput", "6761": "deprecated/deta/modeling_deta.py:DetaObjectDetectionOutput", "6762": "deprecated/deta/modeling_deta.py:_get_clones", "6763": "deprecated/deta/modeling_deta.py:inverse_sigmoid", "6764": "deprecated/deta/modeling_deta.py:DetaFrozenBatchNorm2d", "6765": "deprecated/deta/modeling_deta.py:replace_batch_norm", "6766": "deprecated/deta/modeling_deta.py:DetaBackboneWithPositionalEncodings", "6767": "deprecated/deta/modeling_deta.py:DetaSinePositionEmbedding", "6768": "deprecated/deta/modeling_deta.py:DetaLearnedPositionEmbedding", "6769": "deprecated/deta/modeling_deta.py:build_position_encoding", "6770": "deprecated/deta/modeling_deta.py:multi_scale_deformable_attention", "6771": "deprecated/deta/modeling_deta.py:DetaMultiscaleDeformableAttention", "6772": "deprecated/deta/modeling_deta.py:DetaMultiheadAttention", "6773": "deprecated/deta/modeling_deta.py:DetaEncoderLayer", "6774": "deprecated/deta/modeling_deta.py:DetaDecoderLayer", "6775": "deprecated/deta/modeling_deta.py:DetaPreTrainedModel", "6776": "deprecated/deta/modeling_deta.py:DetaEncoder", "6777": "deprecated/deta/modeling_deta.py:DetaDecoder", "6778": "deprecated/deta/modeling_deta.py:DetaModel", "6779": "deprecated/deta/modeling_deta.py:DetaForObjectDetection", "6780": "deprecated/deta/modeling_deta.py:dice_loss", "6781": "deprecated/deta/modeling_deta.py:sigmoid_focal_loss", "6782": "deprecated/deta/modeling_deta.py:DetaLoss", "6783": "deprecated/deta/modeling_deta.py:DetaMLPPredictionHead", "6784": "deprecated/deta/modeling_deta.py:DetaHungarianMatcher", "6785": "deprecated/deta/modeling_deta.py:_upcast", "6786": "deprecated/deta/modeling_deta.py:box_area", "6787": "deprecated/deta/modeling_deta.py:box_iou", "6788": "deprecated/deta/modeling_deta.py:generalized_box_iou", "6789": "deprecated/deta/modeling_deta.py:nonzero_tuple", "6790": "deprecated/deta/modeling_deta.py:DetaMatcher", "6791": "deprecated/deta/modeling_deta.py:subsample_labels", "6792": "deprecated/deta/modeling_deta.py:sample_topk_per_gt", "6793": "deprecated/deta/modeling_deta.py:DetaStage2Assigner", "6794": "deprecated/deta/modeling_deta.py:DetaStage1Assigner", "6795": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:softmax", "6796": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:ngram_attention_bias", "6797": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:compute_relative_buckets", "6798": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:compute_all_stream_relative_buckets", "6799": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetSeq2SeqLMOutput", "6800": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetSeq2SeqModelOutput", "6801": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetDecoderModelOutput", "6802": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetDecoderLMOutput", "6803": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetPreTrainedModel", "6804": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetPositionalEmbeddings", "6805": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetAttention", "6806": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetFeedForward", "6807": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetNgramSelfAttention", "6808": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetEncoderLayer", "6809": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetDecoderLayer", "6810": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetEncoder", "6811": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetDecoder", "6812": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetModel", "6813": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetForConditionalGeneration", "6814": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetForCausalLM", "6815": "deprecated/xlm_prophetnet/modeling_xlm_prophetnet.py:XLMProphetNetDecoderWrapper", "6816": "deprecated/vit_hybrid/modeling_vit_hybrid.py:ViTHybridEmbeddings", "6817": "deprecated/vit_hybrid/modeling_vit_hybrid.py:ViTHybridPatchEmbeddings", "6818": "deprecated/vit_hybrid/modeling_vit_hybrid.py:ViTHybridSelfAttention", "6819": "deprecated/vit_hybrid/modeling_vit_hybrid.py:ViTHybridSdpaSelfAttention", "6820": "deprecated/vit_hybrid/modeling_vit_hybrid.py:ViTHybridSelfOutput", "6821": "deprecated/vit_hybrid/modeling_vit_hybrid.py:ViTHybridAttention", "6822": "deprecated/vit_hybrid/modeling_vit_hybrid.py:ViTHybridSdpaAttention", "6823": "deprecated/vit_hybrid/modeling_vit_hybrid.py:ViTHybridIntermediate", "6824": "deprecated/vit_hybrid/modeling_vit_hybrid.py:ViTHybridOutput", "6825": "deprecated/vit_hybrid/modeling_vit_hybrid.py:ViTHybridLayer", "6826": "deprecated/vit_hybrid/modeling_vit_hybrid.py:ViTHybridEncoder", "6827": "deprecated/vit_hybrid/modeling_vit_hybrid.py:ViTHybridPreTrainedModel", "6828": "deprecated/vit_hybrid/modeling_vit_hybrid.py:ViTHybridModel", "6829": "deprecated/vit_hybrid/modeling_vit_hybrid.py:ViTHybridPooler", "6830": "deprecated/vit_hybrid/modeling_vit_hybrid.py:ViTHybridForImageClassification", "6831": "deprecated/speech_to_text_2/modeling_speech_to_text_2.py:Speech2Text2SinusoidalPositionalEmbedding", "6832": "deprecated/speech_to_text_2/modeling_speech_to_text_2.py:Speech2Text2Attention", "6833": "deprecated/speech_to_text_2/modeling_speech_to_text_2.py:Speech2Text2DecoderLayer", "6834": "deprecated/speech_to_text_2/modeling_speech_to_text_2.py:Speech2Text2PreTrainedModel", "6835": "deprecated/speech_to_text_2/modeling_speech_to_text_2.py:Speech2Text2Decoder", "6836": "deprecated/speech_to_text_2/modeling_speech_to_text_2.py:Speech2Text2DecoderWrapper", "6837": "deprecated/speech_to_text_2/modeling_speech_to_text_2.py:Speech2Text2ForCausalLM", "6838": "deprecated/jukebox/modeling_jukebox.py:filter_logits", "6839": "deprecated/jukebox/modeling_jukebox.py:get_relevant_lyric_tokens", "6840": "deprecated/jukebox/modeling_jukebox.py:get_starts", "6841": "deprecated/jukebox/modeling_jukebox.py:get_alignment", "6842": "deprecated/jukebox/modeling_jukebox.py:save_temp_audio", "6843": "deprecated/jukebox/modeling_jukebox.py:get_mask", "6844": "deprecated/jukebox/modeling_jukebox.py:JukeboxConv1D", "6845": "deprecated/jukebox/modeling_jukebox.py:JukeboxResConv1DBlock", "6846": "deprecated/jukebox/modeling_jukebox.py:JukeboxResnet1D", "6847": "deprecated/jukebox/modeling_jukebox.py:JukeboxEncoderConvBlock", "6848": "deprecated/jukebox/modeling_jukebox.py:JukeboxEncoder", "6849": "deprecated/jukebox/modeling_jukebox.py:JukeboxDecoderConvBock", "6850": "deprecated/jukebox/modeling_jukebox.py:JukeboxDecoder", "6851": "deprecated/jukebox/modeling_jukebox.py:JukeboxBottleneckBlock", "6852": "deprecated/jukebox/modeling_jukebox.py:JukeboxBottleneck", "6853": "deprecated/jukebox/modeling_jukebox.py:JukeboxVQVAE", "6854": "deprecated/jukebox/modeling_jukebox.py:JukeboxMLP", "6855": "deprecated/jukebox/modeling_jukebox.py:JukeboxLayerNorm", "6856": "deprecated/jukebox/modeling_jukebox.py:JukeboxAttention", "6857": "deprecated/jukebox/modeling_jukebox.py:JukeboxBlock", "6858": "deprecated/jukebox/modeling_jukebox.py:JukeboxLayerStack", "6859": "deprecated/jukebox/modeling_jukebox.py:JukeboxPositionalEmbedding", "6860": "deprecated/jukebox/modeling_jukebox.py:JukeboxConditionalAutoregressive", "6861": "deprecated/jukebox/modeling_jukebox.py:JukeboxMusicTokenConditioner", "6862": "deprecated/jukebox/modeling_jukebox.py:JukeboxRangeEmbedding", "6863": "deprecated/jukebox/modeling_jukebox.py:JukeboxLabelConditioner", "6864": "deprecated/jukebox/modeling_jukebox.py:JukeboxPrior", "6865": "deprecated/jukebox/modeling_jukebox.py:JukeboxPreTrainedModel", "6866": "deprecated/jukebox/modeling_jukebox.py:JukeboxModel", "6867": "deprecated/nat/modeling_nat.py:NatEncoderOutput", "6868": "deprecated/nat/modeling_nat.py:NatModelOutput", "6869": "deprecated/nat/modeling_nat.py:NatImageClassifierOutput", "6870": "deprecated/nat/modeling_nat.py:NatEmbeddings", "6871": "deprecated/nat/modeling_nat.py:NatPatchEmbeddings", "6872": "deprecated/nat/modeling_nat.py:NatDownsampler", "6873": "deprecated/nat/modeling_nat.py:drop_path", "6874": "deprecated/nat/modeling_nat.py:NatDropPath", "6875": "deprecated/nat/modeling_nat.py:NeighborhoodAttention", "6876": "deprecated/nat/modeling_nat.py:NeighborhoodAttentionOutput", "6877": "deprecated/nat/modeling_nat.py:NeighborhoodAttentionModule", "6878": "deprecated/nat/modeling_nat.py:NatIntermediate", "6879": "deprecated/nat/modeling_nat.py:NatOutput", "6880": "deprecated/nat/modeling_nat.py:NatLayer", "6881": "deprecated/nat/modeling_nat.py:NatStage", "6882": "deprecated/nat/modeling_nat.py:NatEncoder", "6883": "deprecated/nat/modeling_nat.py:NatPreTrainedModel", "6884": "deprecated/nat/modeling_nat.py:NatModel", "6885": "deprecated/nat/modeling_nat.py:NatForImageClassification", "6886": "deprecated/nat/modeling_nat.py:NatBackbone", "6887": "deprecated/ernie_m/modeling_ernie_m.py:ErnieMEmbeddings", "6888": "deprecated/ernie_m/modeling_ernie_m.py:ErnieMSelfAttention", "6889": "deprecated/ernie_m/modeling_ernie_m.py:ErnieMAttention", "6890": "deprecated/ernie_m/modeling_ernie_m.py:ErnieMEncoderLayer", "6891": "deprecated/ernie_m/modeling_ernie_m.py:ErnieMEncoder", "6892": "deprecated/ernie_m/modeling_ernie_m.py:ErnieMPooler", "6893": "deprecated/ernie_m/modeling_ernie_m.py:ErnieMPreTrainedModel", "6894": "deprecated/ernie_m/modeling_ernie_m.py:ErnieMModel", "6895": "deprecated/ernie_m/modeling_ernie_m.py:ErnieMForSequenceClassification", "6896": "deprecated/ernie_m/modeling_ernie_m.py:ErnieMForMultipleChoice", "6897": "deprecated/ernie_m/modeling_ernie_m.py:ErnieMForTokenClassification", "6898": "deprecated/ernie_m/modeling_ernie_m.py:ErnieMForQuestionAnswering", "6899": "deprecated/ernie_m/modeling_ernie_m.py:ErnieMForInformationExtraction", "6900": "deprecated/mega/modeling_mega.py:MegaEmbeddings", "6901": "deprecated/mega/modeling_mega.py:MegaSimpleRelativePositionalBias", "6902": "deprecated/mega/modeling_mega.py:MegaRotaryRelativePositionalBias", "6903": "deprecated/mega/modeling_mega.py:MegaDropout", "6904": "deprecated/mega/modeling_mega.py:MegaRMSNorm", "6905": "deprecated/mega/modeling_mega.py:MegaScaleNorm", "6906": "deprecated/mega/modeling_mega.py:MegaSequenceNorm", "6907": "deprecated/mega/modeling_mega.py:MegaMultiDimensionDampedEma", "6908": "deprecated/mega/modeling_mega.py:MegaGatedCrossAttention", "6909": "deprecated/mega/modeling_mega.py:MegaMovingAverageGatedAttention", "6910": "deprecated/mega/modeling_mega.py:MegaNormalizedFeedForwardNetwork", "6911": "deprecated/mega/modeling_mega.py:MegaBlock", "6912": "deprecated/mega/modeling_mega.py:MegaPooler", "6913": "deprecated/mega/modeling_mega.py:MegaPreTrainedModel", "6914": "deprecated/mega/modeling_mega.py:MegaModel", "6915": "deprecated/mega/modeling_mega.py:MegaForCausalLM", "6916": "deprecated/mega/modeling_mega.py:MegaForMaskedLM", "6917": "deprecated/mega/modeling_mega.py:MegaForSequenceClassification", "6918": "deprecated/mega/modeling_mega.py:MegaForMultipleChoice", "6919": "deprecated/mega/modeling_mega.py:MegaForTokenClassification", "6920": "deprecated/mega/modeling_mega.py:MegaClassificationHead", "6921": "deprecated/mega/modeling_mega.py:MegaForQuestionAnswering", "6922": "deprecated/retribert/modeling_retribert.py:RetriBertPreTrainedModel", "6923": "deprecated/retribert/modeling_retribert.py:RetriBertModel", "6924": "deprecated/nezha/modeling_nezha.py:NezhaRelativePositionsEncoding", "6925": "deprecated/nezha/modeling_nezha.py:NezhaEmbeddings", "6926": "deprecated/nezha/modeling_nezha.py:NezhaSelfAttention", "6927": "deprecated/nezha/modeling_nezha.py:NezhaSelfOutput", "6928": "deprecated/nezha/modeling_nezha.py:NezhaAttention", "6929": "deprecated/nezha/modeling_nezha.py:NezhaIntermediate", "6930": "deprecated/nezha/modeling_nezha.py:NezhaOutput", "6931": "deprecated/nezha/modeling_nezha.py:NezhaLayer", "6932": "deprecated/nezha/modeling_nezha.py:NezhaEncoder", "6933": "deprecated/nezha/modeling_nezha.py:NezhaPooler", "6934": "deprecated/nezha/modeling_nezha.py:NezhaPredictionHeadTransform", "6935": "deprecated/nezha/modeling_nezha.py:NezhaLMPredictionHead", "6936": "deprecated/nezha/modeling_nezha.py:NezhaOnlyMLMHead", "6937": "deprecated/nezha/modeling_nezha.py:NezhaOnlyNSPHead", "6938": "deprecated/nezha/modeling_nezha.py:NezhaPreTrainingHeads", "6939": "deprecated/nezha/modeling_nezha.py:NezhaPreTrainedModel", "6940": "deprecated/nezha/modeling_nezha.py:NezhaForPreTrainingOutput", "6941": "deprecated/nezha/modeling_nezha.py:NezhaModel", "6942": "deprecated/nezha/modeling_nezha.py:NezhaForPreTraining", "6943": "deprecated/nezha/modeling_nezha.py:NezhaForMaskedLM", "6944": "deprecated/nezha/modeling_nezha.py:NezhaForNextSentencePrediction", "6945": "deprecated/nezha/modeling_nezha.py:NezhaForSequenceClassification", "6946": "deprecated/nezha/modeling_nezha.py:NezhaForMultipleChoice", "6947": "deprecated/nezha/modeling_nezha.py:NezhaForTokenClassification", "6948": "deprecated/nezha/modeling_nezha.py:NezhaForQuestionAnswering", "6949": "deprecated/mctct/modeling_mctct.py:MCTCTConv1dSubsampler", "6950": "deprecated/mctct/modeling_mctct.py:MCTCTEmbeddings", "6951": "deprecated/mctct/modeling_mctct.py:MCTCTSelfAttention", "6952": "deprecated/mctct/modeling_mctct.py:MCTCTLayerNorm", "6953": "deprecated/mctct/modeling_mctct.py:MCTCTSelfOutput", "6954": "deprecated/mctct/modeling_mctct.py:MCTCTAttention", "6955": "deprecated/mctct/modeling_mctct.py:MCTCTIntermediate", "6956": "deprecated/mctct/modeling_mctct.py:MCTCTOutput", "6957": "deprecated/mctct/modeling_mctct.py:MCTCTLayer", "6958": "deprecated/mctct/modeling_mctct.py:MCTCTPreTrainedModel", "6959": "deprecated/mctct/modeling_mctct.py:MCTCTEncoder", "6960": "deprecated/mctct/modeling_mctct.py:MCTCTModel", "6961": "deprecated/mctct/modeling_mctct.py:MCTCTForCTC", "6962": "deprecated/mmbt/modeling_mmbt.py:ModalEmbeddings", "6963": "deprecated/mmbt/modeling_mmbt.py:MMBTModel", "6964": "deprecated/mmbt/modeling_mmbt.py:MMBTForClassification", "6965": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerPatchEmbeddings", "6966": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerSelfAttention", "6967": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerConvStem", "6968": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerPooling", "6969": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerDenseMlp", "6970": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerConvMlp", "6971": "deprecated/efficientformer/modeling_efficientformer.py:drop_path", "6972": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerDropPath", "6973": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerFlat", "6974": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerMeta3D", "6975": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerMeta3DLayers", "6976": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerMeta4D", "6977": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerMeta4DLayers", "6978": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerIntermediateStage", "6979": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerLastStage", "6980": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerEncoder", "6981": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerPreTrainedModel", "6982": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerModel", "6983": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerForImageClassification", "6984": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerForImageClassificationWithTeacherOutput", "6985": "deprecated/efficientformer/modeling_efficientformer.py:EfficientFormerForImageClassificationWithTeacher", "6986": "deprecated/van/modeling_van.py:drop_path", "6987": "deprecated/van/modeling_van.py:VanDropPath", "6988": "deprecated/van/modeling_van.py:VanOverlappingPatchEmbedder", "6989": "deprecated/van/modeling_van.py:VanMlpLayer", "6990": "deprecated/van/modeling_van.py:VanLargeKernelAttention", "6991": "deprecated/van/modeling_van.py:VanLargeKernelAttentionLayer", "6992": "deprecated/van/modeling_van.py:VanSpatialAttentionLayer", "6993": "deprecated/van/modeling_van.py:VanLayerScaling", "6994": "deprecated/van/modeling_van.py:VanLayer", "6995": "deprecated/van/modeling_van.py:VanStage", "6996": "deprecated/van/modeling_van.py:VanEncoder", "6997": "deprecated/van/modeling_van.py:VanPreTrainedModel", "6998": "deprecated/van/modeling_van.py:VanModel", "6999": "deprecated/van/modeling_van.py:VanForImageClassification", "7000": "deprecated/open_llama/modeling_open_llama.py:OpenLlamaRMSNorm", "7001": "deprecated/open_llama/modeling_open_llama.py:OpenLlamaRotaryEmbedding", "7002": "deprecated/open_llama/modeling_open_llama.py:OpenLlamaLinearScalingRotaryEmbedding", "7003": "deprecated/open_llama/modeling_open_llama.py:OpenLlamaDynamicNTKScalingRotaryEmbedding", "7004": "deprecated/open_llama/modeling_open_llama.py:rotate_half", "7005": "deprecated/open_llama/modeling_open_llama.py:apply_rotary_pos_emb", "7006": "deprecated/open_llama/modeling_open_llama.py:OpenLlamaMLP", "7007": "deprecated/open_llama/modeling_open_llama.py:OpenLlamaAttention", "7008": "deprecated/open_llama/modeling_open_llama.py:OpenLlamaDecoderLayer", "7009": "deprecated/open_llama/modeling_open_llama.py:OpenLlamaPreTrainedModel", "7010": "deprecated/open_llama/modeling_open_llama.py:OpenLlamaModel", "7011": "deprecated/open_llama/modeling_open_llama.py:OpenLlamaForCausalLM", "7012": "deprecated/open_llama/modeling_open_llama.py:OpenLlamaForSequenceClassification", "7013": "deprecated/trajectory_transformer/modeling_trajectory_transformer.py:TrajectoryTransformerOutput", "7014": "deprecated/trajectory_transformer/modeling_trajectory_transformer.py:TrajectoryTransformerPreTrainedModel", "7015": "deprecated/trajectory_transformer/modeling_trajectory_transformer.py:EinLinear", "7016": "deprecated/trajectory_transformer/modeling_trajectory_transformer.py:CausalSelfAttention", "7017": "deprecated/trajectory_transformer/modeling_trajectory_transformer.py:Block", "7018": "deprecated/trajectory_transformer/modeling_trajectory_transformer.py:TrajectoryTransformerModel", "7019": "deprecated/gptsan_japanese/modeling_gptsan_japanese.py:router_z_loss_func", "7020": "deprecated/gptsan_japanese/modeling_gptsan_japanese.py:load_balancing_loss_func", "7021": "deprecated/gptsan_japanese/modeling_gptsan_japanese.py:GPTSanJapaneseDenseActDense", "7022": "deprecated/gptsan_japanese/modeling_gptsan_japanese.py:GPTSanJapaneseTop1Router", "7023": "deprecated/gptsan_japanese/modeling_gptsan_japanese.py:GPTSanJapaneseSparseMLP", "7024": "deprecated/gptsan_japanese/modeling_gptsan_japanese.py:GPTSanJapaneseLayerSparseFF", "7025": "deprecated/gptsan_japanese/modeling_gptsan_japanese.py:GPTSanJapaneseLayerDenseFF", "7026": "deprecated/gptsan_japanese/modeling_gptsan_japanese.py:GPTSanJapaneseAttention", "7027": "deprecated/gptsan_japanese/modeling_gptsan_japanese.py:GPTSanJapaneseLayerSelfAttention", "7028": "deprecated/gptsan_japanese/modeling_gptsan_japanese.py:GPTSanJapaneseBlock", "7029": "deprecated/gptsan_japanese/modeling_gptsan_japanese.py:GPTSanJapanesePreTrainedModel", "7030": "deprecated/gptsan_japanese/modeling_gptsan_japanese.py:GPTSanJapaneseModel", "7031": "deprecated/gptsan_japanese/modeling_gptsan_japanese.py:GPTSanJapaneseForConditionalGeneration", "7032": "deprecated/graphormer/modeling_graphormer.py:quant_noise", "7033": "deprecated/graphormer/modeling_graphormer.py:LayerDropModuleList", "7034": "deprecated/graphormer/modeling_graphormer.py:GraphormerGraphNodeFeature", "7035": "deprecated/graphormer/modeling_graphormer.py:GraphormerGraphAttnBias", "7036": "deprecated/graphormer/modeling_graphormer.py:GraphormerMultiheadAttention", "7037": "deprecated/graphormer/modeling_graphormer.py:GraphormerGraphEncoderLayer", "7038": "deprecated/graphormer/modeling_graphormer.py:GraphormerGraphEncoder", "7039": "deprecated/graphormer/modeling_graphormer.py:GraphormerDecoderHead", "7040": "deprecated/graphormer/modeling_graphormer.py:GraphormerPreTrainedModel", "7041": "deprecated/graphormer/modeling_graphormer.py:GraphormerModel", "7042": "deprecated/graphormer/modeling_graphormer.py:GraphormerForGraphClassification"}