download history blame contribute delete
Detected Pickle imports (110)
- "__torch__.torch.nn.modules.normalization.___torch_mangle_17.LayerNorm",
- "__torch__.PatchTST.FlattenHead",
- "__torch__.torch.nn.modules.linear.___torch_mangle_10.Linear",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_18.LayerNorm",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_31.LayerNorm",
- "__torch__.torch.nn.modules.flatten.Flatten",
- "__torch__.layers.Transformer_EncDec.EncoderLayer",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_21.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_36.Dropout",
- "__torch__.torch.nn.modules.normalization.LayerNorm",
- "__torch__.layers.SelfAttention_Family.FullAttention",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_19.Dropout",
- "__torch__.layers.Embed.PatchEmbedding",
- "__torch__.layers.Transformer_EncDec.Encoder",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_34.LayerNorm",
- "__torch__.layers.Embed.PositionalEmbedding",
- "__torch__.torch.nn.modules.linear.___torch_mangle_26.Linear",
- "collections.OrderedDict",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_32.Dropout",
- "__torch__.layers.SelfAttention_Family.AttentionLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_24.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_23.Linear",
- "__torch__.layers.Transformer_EncDec.___torch_mangle_33.EncoderLayer",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_6.LayerNorm",
- "__torch__.torch.nn.modules.conv.___torch_mangle_29.Conv1d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_12.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_11.Linear",
- "torch.FloatStorage",
- "__torch__.torch.nn.modules.container.ModuleList",
- "__torch__.torch.nn.modules.linear.___torch_mangle_1.Linear",
- "__torch__.layers.SelfAttention_Family.___torch_mangle_9.FullAttention",
- "__torch__.torch.nn.modules.conv.___torch_mangle_16.Conv1d",
- "__torch__.torch.nn.modules.normalization.___torch_mangle_30.LayerNorm",
- "__torch__.layers.Transformer_EncDec.___torch_mangle_20.EncoderLayer",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.torch.nn.modules.linear.___torch_mangle_4.Linear",
- "__torch__.layers.SelfAttention_Family.___torch_mangle_14.AttentionLayer",
- "__torch__.torch.nn.modules.dropout.Dropout",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_8.Dropout",
- "__torch__.torch.nn.modules.padding.ReplicationPad1d",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_7.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_2.Linear",
- "__torch__.torch.nn.modules.dropout.___torch_mangle_0.Dropout",
- "__torch__.torch.nn.modules.linear.___torch_mangle_25.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_15.Conv1d",
- "__torch__.torch.nn.modules.conv.___torch_mangle_5.Conv1d",
- "__torch__.PatchTST.Model",
- "__torch__.torch.nn.modules.conv.Conv1d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_3.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_13.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_28.Conv1d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_35.Linear",
- "__torch__.layers.SelfAttention_Family.___torch_mangle_22.FullAttention",
- "__torch__.torch.nn.modules.linear.Linear",
- "__torch__.layers.SelfAttention_Family.___torch_mangle_27.AttentionLayer",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.jit._pickle.build_intlist",
- "torch.DoubleStorage",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.jit._pickle.restore_type_tag",
- "torch._utils._rebuild_tensor_v2",
- "collections.OrderedDict",
- "torch.FloatStorage"
4.54 MB
- SHA256:
- 1124776f6ca71f3c2759e58c3668778a5752bb03acbddd6d7b1d0479eec407e1
- Pointer size:
- 132 Bytes
- Size of remote file:
- 4.54 MB
- Xet hash:
- 87b02bfedf14c078a80ee9947062d65e1d81a838b67f25c76e4eb04c6988d085
·
·
Xet efficiently stores Large Files inside Git, intelligently splitting files into unique chunks and accelerating uploads and downloads. More info.