From 059c55f8cc9b33dcbf78c71afad20cf640eae070 Mon Sep 17 00:00:00 2001 From: Johannes M Dieterich Date: Tue, 12 Feb 2019 16:45:09 -0800 Subject: [PATCH] Enable test_jit tests that work on ROCm 2.1 Summary: Pull Request resolved: https://github.com/pytorch/pytorch/pull/17010 Differential Revision: D14059748 Pulled By: bddppq fbshipit-source-id: 7a1f7eee4f818dba91e741437415370973e4d429 --- test/test_jit.py | 16 ---------------- 1 file changed, 16 deletions(-) diff --git a/test/test_jit.py b/test/test_jit.py index f2ab4bf..8847b51 100644 --- a/test/test_jit.py +++ b/test/test_jit.py @@ -1425,7 +1425,6 @@ class TestJit(JitTestCase): @unittest.skipIf(IS_WINDOWS, "NYI: fuser support for Windows") @unittest.skipIf(not RUN_CUDA, "requires CUDA") - @skipIfRocm def test_ge_cuda(self): self.run_ge_tests(True, True) @@ -1498,7 +1497,6 @@ class TestJit(JitTestCase): @unittest.skipIf(IS_WINDOWS, "NYI: fuser support for Windows") @unittest.skipIf(not RUN_CUDA, "calls .cuda()") - @skipIfRocm def test_traced_module_cuda(self): class Model(nn.Module): def __init__(self, num_features, num_layers): @@ -4983,7 +4981,6 @@ a") self.assertEqual(b.device, s(b, "t.to(dtype=torch.int32)").device) @unittest.skipIf(not RUN_CUDA, "No CUDA") - @skipIfRocm def test_tensor_number_math_cuda(self): self._test_tensor_number_math(device='cuda') @@ -10036,7 +10033,6 @@ class TestEndToEndHybridFrontendModels(JitTestCase): self._test_dcgan_models(self, device='cpu') @unittest.skipIf(not RUN_CUDA, "no CUDA") - @skipIfRocm def test_dcgan_models_cuda(self): # XXX: export_import on CUDA modules doesn't work (#11480) self._test_dcgan_models(self, device='cuda', check_export_import=False) @@ -10159,13 +10155,11 @@ class TestEndToEndHybridFrontendModels(JitTestCase): self._test_mnist(self, device='cpu') @unittest.skipIf(not RUN_CUDA, "no CUDA") - @skipIfRocm def test_mnist_cuda(self): # XXX: export_import on CUDA modules doesn't work (#11480) self._test_mnist(self, device='cuda', check_export_import=False) @unittest.skipIf(not RUN_CUDA, "no CUDA") - @skipIfRocm def test_mnist_training_leaks_no_memory_cuda(self): net = MnistNet().cuda() # MnistNet uses dropout, don't check its trace @@ -10313,16 +10307,13 @@ class TestEndToEndHybridFrontendModels(JitTestCase): self.checkTrace(SNLIClassifier(Config()).to(device), (premise, hypothesis), inputs_require_grads=False, export_import=check_export_import) - @skipIfRocm def test_snli(self): self._test_snli(self, device='cpu') if not TEST_WITH_UBSAN and torch.fbgemm_is_cpu_supported(): - @skipIfRocm def test_snli_quantized(self): self._test_snli(self, device='cpu', quantized=True) - @skipIfRocm @unittest.skipIf(not RUN_CUDA, "no CUDA") def test_snli_cuda(self): # XXX: export_import on CUDA modules doesn't work (#11480) @@ -10355,11 +10346,9 @@ class TestEndToEndHybridFrontendModels(JitTestCase): self.checkTrace(net, (torch.rand(5, 1, 32, 32, device=device),), export_import=check_export_import) - @skipIfRocm def test_super_resolution(self): self._test_super_resolution(self, device='cpu') - @skipIfRocm @unittest.skipIf(not RUN_CUDA, 'no CUDA') def test_super_resolution_cuda(self): # XXX: export_import on CUDA modules doesn't work (#11480) @@ -10528,7 +10517,6 @@ class TestPytorchExportModes(JitTestCase): x = torch.ones(3) torch.onnx._export(foo, (x,), f, example_outputs=(x, x)) - @skipIfRocm @skipIfNoLapack def test_aten_fallback(self): class ModelWithAtenNotONNXOp(nn.Module): @@ -10548,7 +10536,6 @@ class TestPytorchExportModes(JitTestCase): # torch.fmod is using to test ONNX_ATEN. # If you plan to remove fmod from aten, or found this test failed. # please contact @Rui. - @skipIfRocm def test_onnx_aten(self): class ModelWithAtenFmod(nn.Module): def forward(self, x, y): @@ -10858,7 +10845,6 @@ class TestFuser(JitTestCase): @unittest.skipIf(IS_WINDOWS, "NYI: fuser support for Windows") @unittest.skipIf(not RUN_CUDA, "fuser requires CUDA") - @skipIfRocm def test_arg_configurations_smoke_cuda(self): # A smoke test to make sure we won't use the same kernel for contiguous # and non-contiguous arguments. @@ -10987,14 +10973,12 @@ class TestFuser(JitTestCase): self.checkScript(fn, [tensor]) @unittest.skipIf(IS_WINDOWS or IS_SANDCASTLE, "NYI: fuser support for Windows or Sandcastle") - @skipIfRocm @enable_cpu_fuser def test_chunk_correctness(self): return self._test_chunk_correctness(self, 'cpu') @unittest.skipIf(IS_WINDOWS, "NYI: fuser support for Windows") @unittest.skipIf(not RUN_CUDA, "No CUDA") - @skipIfRocm def test_chunk_correctness_cuda(self): return self._test_chunk_correctness(self, 'cuda') -- 2.7.4