diff --git a/tensorflow/compiler/tests/xla_call_module_test.py b/tensorflow/compiler/tests/xla_call_module_test.py index 238ef0e3b946dd..ea31b4c072e607 100644 --- a/tensorflow/compiler/tests/xla_call_module_test.py +++ b/tensorflow/compiler/tests/xla_call_module_test.py @@ -445,7 +445,10 @@ def f(x): ) self._assertOpOutputMatchesExpected(f, (x,), (expected_value,)) + def test_platforms_and_poly_and_tokens(self): + if test.is_built_with_rocm(): + self.skipTest('Currently failing on ROCm due to mismatch') x = np.arange(6, dtype=np.float32) # returns x + 2. on CPU, x + 3. on GPU (CUDA or ROCM) and x + 4. on TPU @@ -496,7 +499,7 @@ def f(x): } } """ - + def platforms_errors_helper( self, *, diff --git a/tensorflow/python/eager/context_test.py b/tensorflow/python/eager/context_test.py index 9d93e3650b6878..69ca5b0c45a23a 100644 --- a/tensorflow/python/eager/context_test.py +++ b/tensorflow/python/eager/context_test.py @@ -203,6 +203,8 @@ def test_func(x): result = test_func.experimental_get_compiler_ir(a)(stage=stage) self.assertNotEmpty(result) if stage == 'optimized_hlo_proto_serialized': + if test.is_built_with_rocm(): + self.skipTest('Currently failing on ROCm due to mismatch') hlo_proto = hlo_pb2.HloProto.FromString(result) allocations = hlo_proto.buffer_assignment.buffer_allocations buffer_size = sum( diff --git a/tensorflow/python/ops/memory_tests/custom_gradient_memory_test.py b/tensorflow/python/ops/memory_tests/custom_gradient_memory_test.py index c59009fb809e90..e8200615f92969 100644 --- a/tensorflow/python/ops/memory_tests/custom_gradient_memory_test.py +++ b/tensorflow/python/ops/memory_tests/custom_gradient_memory_test.py @@ -112,6 +112,8 @@ def run(test_func): @test_util.run_v2_only def testRecomputeGradXla(self): + if test.is_built_with_rocm(): + self.skipTest('Currently failing on ROCm due to mismatch') device_type = self._get_device_type() device_name = f"{device_type}:0" # Necessary for TFRT tests.