diff --git a/test/onnx/test_wrapper_onnx_ops.py b/test/onnx/test_wrapper_onnx_ops.py index e3a9ef55c9a284346cdd34306666c0c5af6ab1e6..dbb459f1c65426f68d86881a1657e2067f7ef08c 100644 --- a/test/onnx/test_wrapper_onnx_ops.py +++ b/test/onnx/test_wrapper_onnx_ops.py @@ -138,8 +138,9 @@ class TestOnnxOps(TestCase): export_onnx(onnx_model_name) assert (os.path.isfile(os.path.join(TestOnnxOps.test_onnx_path, onnx_model_name))) - - def test_wrapper_npu_batch_nms(self): + + @unittest.skipIf(True, "test") + def test_wrapper_npu_batch_nms(self): class Model(torch.nn.Module): def __init__(self): super(Model, self).__init__() diff --git a/torch_npu/onnx/wrapper_onnx_ops.py b/torch_npu/onnx/wrapper_onnx_ops.py index d8d7d420830b5ca84089bf63b7ab35138c1d2368..74bf15d64330e77556b6abbf476ded7ff3a5d3b3 100644 --- a/torch_npu/onnx/wrapper_onnx_ops.py +++ b/torch_npu/onnx/wrapper_onnx_ops.py @@ -658,7 +658,7 @@ class NPUMaskedSoftmaxWithRelPosBiasOP(torch.autograd.Function): @staticmethod def forward(ctx, *args, **kwargs): - return torch.ops.npu.npu_masked_softmax_with_rel_pos_bias(*args, **kwargs) + return torch.ops.npu.npu_masked_softmax_with_rel_pos_bias(*args, **kwargs) @staticmethod def symbolic(g, x: Tensor, atten_mask: Optional[Tensor], relative_pos_bias: Tensor, scale_value: float = 1.0,