upload files
c23ef24
-
log
upload files
-
tensorboard
upload files
-
443 Bytes
upload files
epoch-40.pt
Detected Pickle imports (4)
- "collections.OrderedDict",
- "torch.FloatStorage",
- "pathlib.PosixPath",
- "torch._utils._rebuild_tensor_v2"
How to fix it?
373 MB
upload files
-
320 Bytes
upload files
-
276 Bytes
upload files
jit_script.pt
Detected Pickle imports (98)
- "__torch__.scaling.___torch_mangle_5.ScheduledFloat",
- "__torch__.scaling.SwooshL",
- "__torch__.torch.nn.modules.conv.___torch_mangle_2.Conv2d",
- "torch._utils._rebuild_tensor_v2",
- "__torch__.torch.nn.modules.linear.___torch_mangle_21.Linear",
- "__torch__.zipformer.___torch_mangle_31.NonlinAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_56.Linear",
- "__torch__.subsampling.ConvNeXt",
- "__torch__.torch.nn.modules.linear.___torch_mangle_10.Linear",
- "__torch__.zipformer.BypassModule",
- "__torch__.zipformer.___torch_mangle_40.Zipformer2EncoderLayer",
- "__torch__.torch.nn.modules.container.Sequential",
- "torch.jit._pickle.build_intlist",
- "__torch__.scaling.SwooshR",
- "__torch__.torch.nn.modules.linear.___torch_mangle_30.Linear",
- "__torch__.zipformer.ConvolutionModule",
- "__torch__.torch.nn.modules.container.___torch_mangle_36.ModuleList",
- "__torch__.zipformer.___torch_mangle_53.DownsampledZipformer2Encoder",
- "__torch__.torch.nn.modules.conv.___torch_mangle_0.Conv2d",
- "__torch__.decoder.Decoder",
- "__torch__.zipformer.___torch_mangle_42.Zipformer2Encoder",
- "__torch__.torch.nn.modules.linear.___torch_mangle_8.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_3.Conv2d",
- "__torch__.torch.nn.modules.conv.___torch_mangle_4.Conv2d",
- "__torch__.zipformer.FeedforwardModule",
- "__torch__.zipformer.___torch_mangle_13.FeedforwardModule",
- "__torch__.torch.nn.modules.conv.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_11.Linear",
- "__torch__.scaling.BiasNorm",
- "__torch__.model.Transducer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_57.Linear",
- "__torch__.scaling.ScheduledFloat",
- "__torch__.scaling.Identity",
- "__torch__.torch.nn.modules.container.___torch_mangle_41.ModuleList",
- "collections.OrderedDict",
- "__torch__.torch.nn.modules.linear.___torch_mangle_9.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_14.Linear",
- "__torch__.EncoderModel",
- "__torch__.zipformer.Zipformer2",
- "__torch__.torch.nn.modules.conv.___torch_mangle_55.Conv1d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_32.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_33.Conv1d",
- "__torch__.zipformer.___torch_mangle_43.DownsampledZipformer2Encoder",
- "__torch__.torch.nn.modules.container.ModuleList",
- "__torch__.torch.nn.modules.container.___torch_mangle_51.ModuleList",
- "__torch__.scaling.ActivationDropoutAndLinear",
- "__torch__.torch.nn.modules.activation.Sigmoid",
- "__torch__.zipformer.___torch_mangle_6.BypassModule",
- "__torch__.torch.nn.modules.linear.___torch_mangle_22.Linear",
- "__torch__.zipformer.___torch_mangle_29.FeedforwardModule",
- "__torch__.zipformer.___torch_mangle_20.RelPositionMultiheadAttentionWeights",
- "__torch__.zipformer.Zipformer2EncoderLayer",
- "torch.FloatStorage",
- "__torch__.torch.nn.modules.container.___torch_mangle_54.ModuleList",
- "__torch__.torch.nn.modules.conv.Conv1d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_44.Linear",
- "__torch__.zipformer.SimpleDownsample",
- "__torch__.zipformer.___torch_mangle_52.Zipformer2Encoder",
- "__torch__.torch.nn.modules.linear.___torch_mangle_45.Linear",
- "__torch__.torch.nn.modules.linear.Identity",
- "__torch__.zipformer.___torch_mangle_37.Zipformer2Encoder",
- "__torch__.zipformer.SimpleUpsample",
- "__torch__.zipformer.RelPositionMultiheadAttentionWeights",
- "__torch__.scaling.___torch_mangle_18.ActivationDropoutAndLinear",
- "__torch__.zipformer.Zipformer2Encoder",
- "__torch__.torch.nn.modules.linear.___torch_mangle_48.Linear",
- "__torch__.zipformer.___torch_mangle_34.ConvolutionModule",
- "__torch__.scaling.Dropout2",
- "__torch__.torch.nn.modules.activation.Tanh",
- "__torch__.zipformer.___torch_mangle_25.FeedforwardModule",
- "__torch__.subsampling.Conv2dSubsampling",
- "__torch__.zipformer.SelfAttention",
- "__torch__.zipformer.CompactRelPositionalEncoding",
- "__torch__.zipformer.___torch_mangle_46.RelPositionMultiheadAttentionWeights",
- "__torch__.zipformer.DownsampledZipformer2Encoder",
- "__torch__.torch.nn.modules.sparse.Embedding",
- "__torch__.joiner.Joiner",
- "__torch__.torch.nn.modules.linear.___torch_mangle_26.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_24.Linear",
- "__torch__.zipformer.___torch_mangle_23.SelfAttention",
- "__torch__.torch.nn.modules.linear.___torch_mangle_7.Linear",
- "__torch__.zipformer.___torch_mangle_49.SelfAttention",
- "__torch__.torch.nn.modules.conv.___torch_mangle_38.Conv1d",
- "__torch__.zipformer.___torch_mangle_50.Zipformer2EncoderLayer",
- "__torch__.torch.nn.modules.linear.___torch_mangle_58.Linear",
- "__torch__.torch.nn.modules.conv.___torch_mangle_1.Conv2d",
- "__torch__.torch.nn.modules.linear.___torch_mangle_19.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_47.Linear",
- "__torch__.zipformer.NonlinAttention",
- "__torch__.zipformer.___torch_mangle_35.Zipformer2EncoderLayer",
- "__torch__.zipformer.___torch_mangle_39.ConvolutionModule",
- "__torch__.zipformer.___torch_mangle_15.FeedforwardModule",
- "__torch__.torch.nn.modules.linear.___torch_mangle_16.Linear",
- "__torch__.torch.nn.modules.linear.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_17.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_28.Linear",
- "__torch__.torch.nn.modules.linear.___torch_mangle_12.Linear",
- "__torch__.zipformer.___torch_mangle_27.FeedforwardModule"
How to fix it?
95.8 MB
upload files
-
93.5 MB
upload files
-
467 Bytes
upload files
-
419 Bytes
upload files