operator name
stringclasses
180 values
used in model
stringclasses
155 values
args
stringlengths
19
5.24k
aten.add.Tensor
TorchBench/hf_BigBird
((T([2, 12, 1024, 64], f16), T([2, 12, 1024, 64], f16)), {})
aten.add.Tensor
TorchBench/hf_BigBird
((T([2, 12, 1024, 64], f16, stride=(786432, 65536, 1, 1024)), T([2, 12, 1024, 64], f16)), {})
aten.add.Tensor
TorchBench/hf_BigBird
((T([2, 12, 1024, 64], f16, stride=(786432, 65536, 1, 1024)), T([2, 12, 1024, 64], f16, stride=(786432, 65536, 1, 1024))), {})
aten.add.Tensor
TorchBench/hf_BigBird
((T([2, 12, 12, 64, 512], f16), T([2, 12, 12, 64, 512], f16)), {})
aten.add.Tensor
TorchBench/hf_BigBird
((T([2, 12, 12, 64, 64], f16), T([2, 12, 12, 64, 64], f16)), {})
aten.add.Tensor
TorchBench/hf_BigBird
((T([2, 12, 14, 192, 64], f16), T([2, 12, 14, 192, 64], f16)), {})
aten.add.Tensor
TorchBench/hf_BigBird
((T([2, 12, 16, 64, 64], f16), T([2, 12, 16, 64, 64], f16)), {})
aten.add.Tensor
HuggingFace/M2M100ForConditionalGeneration
((T([2, 128, 1024], f16), T([2, 128, 1024], f16)), {})
aten.add.Tensor
HuggingFace/MegatronBertForCausalLM
((T([2, 128, 1024], f16), T([2, 128, 1024], f16)), {})
aten.add.Tensor
HuggingFace/XGLMForCausalLM
((T([2, 128, 1024], f16), T([2, 128, 1024], f16)), {})
aten.add.Tensor
HuggingFace/M2M100ForConditionalGeneration
((T([2, 128], i32), 0), {})
aten.add.Tensor
HuggingFace/XGLMForCausalLM
((T([2, 128], i32), 0), {})
aten.add.Tensor
HuggingFace/M2M100ForConditionalGeneration
((T([2, 128], i64), 1), {})
aten.add.Tensor
HuggingFace/XGLMForCausalLM
((T([2, 128], i64), 1), {})
aten.add.Tensor
HuggingFace/BartForConditionalGeneration
((T([2, 16, 1024, 1024], f16), T([2, 1, 1024, 1024], f16)), {})
aten.add.Tensor
HuggingFace/MegatronBertForCausalLM
((T([2, 16, 128, 128], f16), T([2, 1, 1, 128], f16)), {})
aten.add.Tensor
HuggingFace/M2M100ForConditionalGeneration
((T([2, 16, 128, 128], f16), T([2, 1, 128, 128], f16)), {})
aten.add.Tensor
HuggingFace/XGLMForCausalLM
((T([2, 16, 128, 128], f16), T([2, 1, 128, 128], f16)), {})
aten.add.Tensor
TIMM/twins_pcpvt_base
((T([2, 32, 1, 49, 64], f16), T([2, 32, 1, 49, 64], f16)), {})
aten.add.Tensor
TIMM/twins_pcpvt_base
((T([2, 32, 2, 49, 64], f16), T([2, 32, 2, 49, 64], f16)), {})
aten.add.Tensor
TIMM/twins_pcpvt_base
((T([2, 32, 5, 49, 64], f16), T([2, 32, 5, 49, 64], f16)), {})
aten.add.Tensor
TIMM/twins_pcpvt_base
((T([2, 32, 8, 49, 64], f16), T([2, 32, 8, 49, 64], f16)), {})
aten.add.Tensor
HuggingFace/AlbertForMaskedLM
((T([2, 512, 128], f16), 1.0), {})
aten.add.Tensor
HuggingFace/AlbertForMaskedLM
((T([2, 512, 128], f16), T([2, 512, 128], f16)), {})
aten.add.Tensor
HuggingFace/AlbertForQuestionAnswering
((T([2, 512, 128], f16), T([2, 512, 128], f16)), {})
aten.add.Tensor
HuggingFace/AlbertForMaskedLM
((T([2, 512, 16384], f16), 1.0), {})
aten.add.Tensor
HuggingFace/AlbertForQuestionAnswering
((T([2, 512, 16384], f16), 1.0), {})
aten.add.Tensor
HuggingFace/AlbertForMaskedLM
((T([2, 512, 16384], f16), T([2, 512, 16384], f16)), {})
aten.add.Tensor
HuggingFace/AlbertForQuestionAnswering
((T([2, 512, 16384], f16), T([2, 512, 16384], f16)), {})
aten.add.Tensor
HuggingFace/AlbertForMaskedLM
((T([2, 512, 4096], f16), T([2, 512, 4096], f16)), {})
aten.add.Tensor
HuggingFace/AlbertForQuestionAnswering
((T([2, 512, 4096], f16), T([2, 512, 4096], f16)), {})
aten.add.Tensor
TIMM/cait_m36_384
((T([2, 576, 576, 16], f16), T([16], f16)), {})
aten.add.Tensor
TIMM/cait_m36_384
((T([2, 576, 768], f16), T([2, 576, 768], f16)), {})
aten.add.Tensor
TIMM/cait_m36_384
((T([2, 576, 768], f16), T([2, 576, 768], f16, stride=(443136, 768, 1))), {})
aten.add.Tensor
TIMM/cait_m36_384
((T([2, 576, 768], f16, stride=(442368, 1, 576)), T([1, 576, 768], f16)), {})
aten.add.Tensor
TIMM/cait_m36_384
((T([2, 576, 768], f16, stride=(442368, 1, 576)), T([2, 576, 768], f16)), {})
aten.add.Tensor
TIMM/cait_m36_384
((T([2, 576, 768], f16, stride=(443136, 768, 1)), T([2, 576, 768], f16, stride=(443136, 768, 1))), {})
aten.add.Tensor
TIMM/cait_m36_384
((T([2, 577, 768], f16), T([2, 577, 768], f16)), {})
aten.add.Tensor
HuggingFace/AlbertForMaskedLM
((T([2, 64, 512, 512], f16), T([2, 1, 1, 512], f16)), {})
aten.add.Tensor
HuggingFace/AlbertForQuestionAnswering
((T([2, 64, 512, 512], f16), T([2, 1, 1, 512], f16)), {})
aten.add.Tensor
TorchBench/vision_maskrcnn
((T([22496, 1, 4], i32), T([1, 3, 4], f16)), {})
aten.add.Tensor
TIMM/coat_lite_mini
((T([24, 1, 5, 5], f16), T([24, 1, 5, 5], f16)), {})
aten.add.Tensor
TIMM/coat_lite_mini
((T([24, 1, 7, 7], f16), T([24, 1, 7, 7], f16)), {})
aten.add.Tensor
TorchBench/hf_Longformer
((T([24, 3, 512, 513], f16), T([24, 3, 512, 513], f16)), {})
aten.add.Tensor
TIMM/coat_lite_mini
((T([24], f16), T([24], f16)), {})
aten.add.Tensor
TIMM/sebotnet33ts_256
((T([256, 1024, 1024], f16), T([256, 1024, 1024], f16)), {})
aten.add.Tensor
TIMM/sebotnet33ts_256
((T([256, 1024, 32], f16), T([256, 1024, 32], f16)), {})
aten.add.Tensor
TIMM/sebotnet33ts_256
((T([256, 16, 16, 128], f16, stride=(32768, 128, 2048, 1)), T([256, 16, 16, 128], f16)), {})
aten.add.Tensor
TIMM/sebotnet33ts_256
((T([256, 16, 16, 16, 16], f16, stride=(8432, 31, 527, 1, 0)), T([256, 16, 16, 16, 16], f16, stride=(8432, 527, 31, 0, 1))), {})
aten.add.Tensor
TIMM/sebotnet33ts_256
((T([256, 16, 16, 64], f16, stride=(16384, 64, 1024, 1)), T([256, 16, 16, 64], f16)), {})
aten.add.Tensor
TIMM/mobilevit_s
((T([256, 16, 240], f16), T([256, 16, 240], f16)), {})
aten.add.Tensor
TIMM/swin_base_patch4_window7_224
((T([256, 16, 49, 49], f16), T([1, 16, 49, 49], f16)), {})
aten.add.Tensor
TIMM/sebotnet33ts_256
((T([256, 256, 128], f16), T([256, 256, 128], f16)), {})
aten.add.Tensor
TIMM/mobilevit_s
((T([256, 256, 144], f16), T([256, 256, 144], f16)), {})
aten.add.Tensor
TIMM/sebotnet33ts_256
((T([256, 256, 256], f16), T([256, 256, 256], f16)), {})
aten.add.Tensor
TIMM/sebotnet33ts_256
((T([256, 256, 64], f16), T([256, 256, 64], f16)), {})
aten.add.Tensor
TorchBench/attention_is_all_you_need_pytorch
((T([256, 31, 512], f16), T([1, 31, 512], f16)), {})
aten.add.Tensor
TorchBench/attention_is_all_you_need_pytorch
((T([256, 31, 512], f16), T([256, 31, 512], f16)), {})
aten.add.Tensor
TIMM/sebotnet33ts_256
((T([256, 32, 32, 32, 32], f16, stride=(66528, 63, 2079, 1, 0)), T([256, 32, 32, 32, 32], f16, stride=(66528, 2079, 63, 0, 1))), {})
aten.add.Tensor
TIMM/sebotnet33ts_256
((T([256, 32, 32, 32], f16, stride=(32768, 32, 1024, 1)), T([256, 32, 32, 32], f16)), {})
aten.add.Tensor
TorchBench/attention_is_all_you_need_pytorch
((T([256, 33, 512], f16), T([1, 33, 512], f16)), {})
aten.add.Tensor
TorchBench/attention_is_all_you_need_pytorch
((T([256, 33, 512], f16), T([256, 33, 512], f16)), {})
aten.add.Tensor
TIMM/sebotnet33ts_256
((T([256, 64, 128], f16), T([256, 64, 128], f16)), {})
aten.add.Tensor
TIMM/mobilevit_s
((T([256, 64, 192], f16), T([256, 64, 192], f16)), {})
aten.add.Tensor
TIMM/sebotnet33ts_256
((T([256, 64, 64], f16), T([256, 64, 64], f16)), {})
aten.add.Tensor
TIMM/sebotnet33ts_256
((T([256, 8, 8, 128], f16, stride=(8192, 128, 1024, 1)), T([256, 8, 8, 128], f16)), {})
aten.add.Tensor
TIMM/sebotnet33ts_256
((T([256, 8, 8, 8, 8], f16, stride=(1080, 15, 135, 1, 0)), T([256, 8, 8, 8, 8], f16, stride=(1080, 135, 15, 0, 1))), {})
aten.add.Tensor
HuggingFace/XGLMForCausalLM
((T([256008, 1024], f16), T([256008, 1024], f16)), {})
aten.add.Tensor
HuggingFace/MegatronBertForCausalLM
((T([29056, 1024], f16), T([29056, 1024], f16)), {})
aten.add.Tensor
TorchBench/timm_efficientdet
((T([2], f16), T([2], f16)), {})
aten.add.Tensor
TorchBench/timm_efficientdet
((T([2], f16), T([2], f16, stride=(0,))), {})
aten.add.Tensor
TorchBench/Background_Matting
((T([3, 128, 256, 256], f16, stride=(16777216, 65536, 256, 1)), T([3, 128, 256, 256], f16)), {})
aten.add.Tensor
TIMM/coat_lite_mini
((T([3, 128, 8, 197, 40], f16), T([3, 128, 8, 197, 40], f16)), {})
aten.add.Tensor
TIMM/coat_lite_mini
((T([3, 128, 8, 3137, 8], f16), T([3, 128, 8, 3137, 8], f16)), {})
aten.add.Tensor
TIMM/coat_lite_mini
((T([3, 128, 8, 50, 64], f16), T([3, 128, 8, 50, 64], f16)), {})
aten.add.Tensor
TIMM/coat_lite_mini
((T([3, 128, 8, 785, 16], f16), T([3, 128, 8, 785, 16], f16)), {})
aten.add.Tensor
TIMM/cait_m36_384
((T([3, 2, 16, 576, 48], f16), T([3, 2, 16, 576, 48], f16)), {})
aten.add.Tensor
TorchBench/Background_Matting
((T([3, 256, 128, 128], f16), T([3, 256, 128, 128], f16)), {})
aten.add.Tensor
TorchBench/Background_Matting
((T([3, 256, 128, 128], f16), T([3, 256, 128, 128], f16, stride=(8388608, 16384, 128, 1))), {})
aten.add.Tensor
TorchBench/Background_Matting
((T([3, 256, 128, 128], f16, stride=(7340032, 16384, 128, 1)), T([3, 256, 128, 128], f16, stride=(8388608, 16384, 128, 1))), {})
aten.add.Tensor
TorchBench/Background_Matting
((T([3, 256, 128, 128], f16, stride=(8388608, 16384, 128, 1)), T([3, 256, 128, 128], f16, stride=(8388608, 16384, 128, 1))), {})
aten.add.Tensor
TorchBench/pytorch_struct
((T([30, 256], f16), T([30, 256], f16)), {})
aten.add.Tensor
TorchBench/pytorch_struct
((T([30, 256], f16, stride=(1, 30)), T([30, 256], f16)), {})
aten.add.Tensor
HuggingFace/AlbertForMaskedLM
((T([30000, 128], f16), T([30000, 128], f16)), {})
aten.add.Tensor
TorchBench/hf_Albert
((T([30000, 128], f16), T([30000, 128], f16)), {})
aten.add.Tensor
HuggingFace/ElectraForCausalLM
((T([30522, 128], f16), T([30522, 128], f16)), {})
aten.add.Tensor
HuggingFace/MobileBertForMaskedLM
((T([30522, 128], f16, stride=(1, 30522)), T([30522, 128], f16)), {})
aten.add.Tensor
HuggingFace/BertForMaskedLM
((T([30522, 768], f16), T([30522, 768], f16)), {})
aten.add.Tensor
HuggingFace/DistilBertForMaskedLM
((T([30522, 768], f16), T([30522, 768], f16)), {})
aten.add.Tensor
HuggingFace/LayoutLMForMaskedLM
((T([30522, 768], f16), T([30522, 768], f16)), {})
aten.add.Tensor
HuggingFace/RobertaForCausalLM
((T([30522, 768], f16), T([30522, 768], f16)), {})
aten.add.Tensor
HuggingFace/YituTechConvBert
((T([30522, 768], f16), T([30522, 768], f16)), {})
aten.add.Tensor
TorchBench/hf_Bert
((T([30522, 768], f16), T([30522, 768], f16)), {})
aten.add.Tensor
TorchBench/hf_DistilBert
((T([30522, 768], f16), T([30522, 768], f16)), {})
aten.add.Tensor
TorchBench/hf_Albert
((T([3072, 768], f16), T([3072, 768], f16)), {})
aten.add.Tensor
TorchBench/hf_Albert
((T([3072], f16), T([3072], f16)), {})
aten.add.Tensor
TIMM/convnext_base
((T([32, 1, 14, 14], f16), 1e-06), {})
aten.add.Tensor
TIMM/convnext_base
((T([32, 1, 28, 28], f16), 1e-06), {})
aten.add.Tensor
TIMM/coat_lite_mini
((T([32, 1, 3, 3], f16), T([32, 1, 3, 3], f16)), {})
aten.add.Tensor
TIMM/convnext_base
((T([32, 1, 56, 56], f16), 1e-06), {})