FEA-Bench / testbed /Project-MONAI__MONAI /tests /test_affine_transform.py
hc99's picture
Add files using upload-large-folder tool
e4b9a7b verified
# Copyright 2020 MONAI Consortium
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
# http://www.apache.org/licenses/LICENSE-2.0
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import unittest
import numpy as np
import torch
from parameterized import parameterized
from monai.networks import normalize_transform, to_norm_affine
from monai.networks.layers import AffineTransform
TEST_NORM_CASES = [
[(4, 5), True, [[[0.666667, 0, -1], [0, 0.5, -1], [0, 0, 1]]]],
[
(2, 4, 5),
True,
[[[2.0, 0.0, 0.0, -1.0], [0.0, 0.6666667, 0.0, -1.0], [0.0, 0.0, 0.5, -1.0], [0.0, 0.0, 0.0, 1.0]]],
],
[(4, 5), False, [[[0.5, 0.0, -0.75], [0.0, 0.4, -0.8], [0.0, 0.0, 1.0]]]],
[(2, 4, 5), False, [[[1.0, 0.0, 0.0, -0.5], [0.0, 0.5, 0.0, -0.75], [0.0, 0.0, 0.4, -0.8], [0.0, 0.0, 0.0, 1.0]]]],
]
TEST_TO_NORM_AFFINE_CASES = [
[
[[[1, 0, 0], [0, 1, 0], [0, 0, 1]]],
(4, 6),
(5, 3),
True,
[[[1.3333334, 0.0, 0.33333337], [0.0, 0.4, -0.6], [0.0, 0.0, 1.0]]],
],
[
[[[1, 0, 0], [0, 1, 0], [0, 0, 1]]],
(4, 6),
(5, 3),
False,
[[[1.25, 0.0, 0.25], [0.0, 0.5, -0.5], [0.0, 0.0, 1.0]]],
],
[
[[[1, 0, 0, 0], [0, 1, 0, 0], [0, 0, 1, 0], [0, 0, 0, 1]]],
(2, 4, 6),
(3, 5, 3),
True,
[[[2.0, 0.0, 0.0, 1.0], [0.0, 1.3333334, 0.0, 0.33333337], [0.0, 0.0, 0.4, -0.6], [0.0, 0.0, 0.0, 1.0]]],
],
[
[[[1, 0, 0, 0], [0, 1, 0, 0], [0, 0, 1, 0], [0, 0, 0, 1]]],
(2, 4, 6),
(3, 5, 3),
False,
[[[1.5, 0.0, 0.0, 0.5], [0.0, 1.25, 0.0, 0.25], [0.0, 0.0, 0.5, -0.5], [0.0, 0.0, 0.0, 1.0]]],
],
]
TEST_ILL_TO_NORM_AFFINE_CASES = [
[[[[1, 0, 0], [0, 1, 0], [0, 0, 1]]], (3, 4, 6), (3, 5, 3), False],
[[[[1, 0, 0, 0], [0, 1, 0, 0], [0, 0, 1, 0], [0, 0, 0, 1]]], (4, 6), (3, 5, 3), True],
[[[[1, 0, 0, 0], [0, 1, 0, 0], [0, 0, 1, 0]]], (4, 6), (3, 5, 3), True],
]
class TestNormTransform(unittest.TestCase):
@parameterized.expand(TEST_NORM_CASES)
def test_norm_xform(self, input_shape, align_corners, expected):
norm = normalize_transform(
input_shape, device=torch.device("cpu:0"), dtype=torch.float32, align_corners=align_corners
)
norm = norm.detach().cpu().numpy()
np.testing.assert_allclose(norm, expected, atol=1e-6)
if torch.cuda.is_available():
norm = normalize_transform(
input_shape, device=torch.device("cuda:0"), dtype=torch.float32, align_corners=align_corners
)
norm = norm.detach().cpu().numpy()
np.testing.assert_allclose(norm, expected, atol=1e-4)
class TestToNormAffine(unittest.TestCase):
@parameterized.expand(TEST_TO_NORM_AFFINE_CASES)
def test_to_norm_affine(self, affine, src_size, dst_size, align_corners, expected):
affine = torch.as_tensor(affine, device=torch.device("cpu:0"), dtype=torch.float32)
new_affine = to_norm_affine(affine, src_size, dst_size, align_corners)
new_affine = new_affine.detach().cpu().numpy()
np.testing.assert_allclose(new_affine, expected, atol=1e-6)
if torch.cuda.is_available():
affine = torch.as_tensor(affine, device=torch.device("cuda:0"), dtype=torch.float32)
new_affine = to_norm_affine(affine, src_size, dst_size, align_corners)
new_affine = new_affine.detach().cpu().numpy()
np.testing.assert_allclose(new_affine, expected, atol=1e-4)
@parameterized.expand(TEST_ILL_TO_NORM_AFFINE_CASES)
def test_to_norm_affine_ill(self, affine, src_size, dst_size, align_corners):
with self.assertRaises(TypeError):
to_norm_affine(affine, src_size, dst_size, align_corners)
with self.assertRaises(ValueError):
affine = torch.as_tensor(affine, device=torch.device("cpu:0"), dtype=torch.float32)
to_norm_affine(affine, src_size, dst_size, align_corners)
class TestAffineTransform(unittest.TestCase):
def test_affine_shift(self):
affine = torch.as_tensor([[1.0, 0.0, 0.0], [0.0, 1.0, -1.0]])
image = torch.as_tensor([[[[4.0, 1.0, 3.0, 2.0], [7.0, 6.0, 8.0, 5.0], [3.0, 5.0, 3.0, 6.0]]]])
out = AffineTransform()(image, affine)
out = out.detach().cpu().numpy()
expected = [[[[0, 4, 1, 3], [0, 7, 6, 8], [0, 3, 5, 3]]]]
np.testing.assert_allclose(out, expected, atol=1e-5)
def test_affine_shift_1(self):
affine = torch.as_tensor([[1.0, 0.0, -1.0], [0.0, 1.0, -1.0]])
image = torch.as_tensor([[[[4.0, 1.0, 3.0, 2.0], [7.0, 6.0, 8.0, 5.0], [3.0, 5.0, 3.0, 6.0]]]])
out = AffineTransform()(image, affine)
out = out.detach().cpu().numpy()
expected = [[[[0, 0, 0, 0], [0, 4, 1, 3], [0, 7, 6, 8]]]]
np.testing.assert_allclose(out, expected, atol=1e-5)
def test_affine_shift_2(self):
affine = torch.as_tensor([[1.0, 0.0, -1.0], [0.0, 1.0, 0.0]])
image = torch.as_tensor([[[[4.0, 1.0, 3.0, 2.0], [7.0, 6.0, 8.0, 5.0], [3.0, 5.0, 3.0, 6.0]]]])
out = AffineTransform()(image, affine)
out = out.detach().cpu().numpy()
expected = [[[[0, 0, 0, 0], [4, 1, 3, 2], [7, 6, 8, 5]]]]
np.testing.assert_allclose(out, expected, atol=1e-5)
def test_zoom(self):
affine = torch.as_tensor([[1.0, 0.0, 0.0], [0.0, 2.0, 0.0]])
image = torch.arange(1.0, 13.0).view(1, 1, 3, 4).to(device=torch.device("cpu:0"))
out = AffineTransform((3, 2))(image, affine)
expected = [[[[1, 3], [5, 7], [9, 11]]]]
np.testing.assert_allclose(out, expected, atol=1e-5)
def test_zoom_1(self):
affine = torch.as_tensor([[2.0, 0.0, 0.0], [0.0, 1.0, 0.0]])
image = torch.arange(1.0, 13.0).view(1, 1, 3, 4).to(device=torch.device("cpu:0"))
out = AffineTransform()(image, affine, (1, 4))
expected = [[[[1, 2, 3, 4]]]]
np.testing.assert_allclose(out, expected, atol=1e-5)
def test_zoom_2(self):
affine = torch.as_tensor([[2.0, 0.0, 0.0], [0.0, 2.0, 0.0]], dtype=torch.float32)
image = torch.arange(1.0, 13.0).view(1, 1, 3, 4).to(device=torch.device("cpu:0"))
out = AffineTransform((1, 2))(image, affine)
expected = [[[[1, 3]]]]
np.testing.assert_allclose(out, expected, atol=1e-5)
def test_affine_transform_minimum(self):
t = np.pi / 3
affine = [[np.cos(t), -np.sin(t), 0], [np.sin(t), np.cos(t), 0], [0, 0, 1]]
affine = torch.as_tensor(affine, device=torch.device("cpu:0"), dtype=torch.float32)
image = torch.arange(24.0).view(1, 1, 4, 6).to(device=torch.device("cpu:0"))
out = AffineTransform()(image, affine)
out = out.detach().cpu().numpy()
expected = [
[
[
[0.0, 0.06698727, 0.0, 0.0, 0.0, 0.0],
[3.8660254, 0.86602557, 0.0, 0.0, 0.0, 0.0],
[7.732051, 3.035899, 0.73205125, 0.0, 0.0, 0.0],
[11.598076, 6.901923, 2.7631402, 0.0, 0.0, 0.0],
]
]
]
np.testing.assert_allclose(out, expected, atol=1e-5)
def test_affine_transform_2d(self):
t = np.pi / 3
affine = [[np.cos(t), -np.sin(t), 0], [np.sin(t), np.cos(t), 0], [0, 0, 1]]
affine = torch.as_tensor(affine, device=torch.device("cpu:0"), dtype=torch.float32)
image = torch.arange(24.0).view(1, 1, 4, 6).to(device=torch.device("cpu:0"))
xform = AffineTransform((3, 4), padding_mode="border", align_corners=True, mode="bilinear")
out = xform(image, affine)
out = out.detach().cpu().numpy()
expected = [
[
[
[7.1525574e-07, 4.9999994e-01, 1.0000000e00, 1.4999999e00],
[3.8660259e00, 1.3660253e00, 1.8660252e00, 2.3660252e00],
[7.7320518e00, 3.0358994e00, 2.7320509e00, 3.2320507e00],
]
]
]
np.testing.assert_allclose(out, expected, atol=1e-5)
if torch.cuda.is_available():
affine = torch.as_tensor(affine, device=torch.device("cuda:0"), dtype=torch.float32)
image = torch.arange(24.0).view(1, 1, 4, 6).to(device=torch.device("cuda:0"))
xform = AffineTransform(padding_mode="border", align_corners=True, mode="bilinear")
out = xform(image, affine, (3, 4))
out = out.detach().cpu().numpy()
expected = [
[
[
[7.1525574e-07, 4.9999994e-01, 1.0000000e00, 1.4999999e00],
[3.8660259e00, 1.3660253e00, 1.8660252e00, 2.3660252e00],
[7.7320518e00, 3.0358994e00, 2.7320509e00, 3.2320507e00],
]
]
]
np.testing.assert_allclose(out, expected, atol=1e-4)
def test_affine_transform_3d(self):
t = np.pi / 3
affine = [[1, 0, 0, 0], [0.0, np.cos(t), -np.sin(t), 0], [0, np.sin(t), np.cos(t), 0], [0, 0, 0, 1]]
affine = torch.as_tensor(affine, device=torch.device("cpu:0"), dtype=torch.float32)
image = torch.arange(48.0).view(2, 1, 4, 2, 3).to(device=torch.device("cpu:0"))
xform = AffineTransform((3, 4, 2), padding_mode="border", align_corners=False, mode="bilinear")
out = xform(image, affine)
out = out.detach().cpu().numpy()
expected = [
[
[
[[0.00000006, 0.5000001], [2.3660254, 1.3660254], [4.732051, 2.4019241], [5.0, 3.9019237]],
[[6.0, 6.5], [8.366026, 7.3660254], [10.732051, 8.401924], [11.0, 9.901924]],
[[12.0, 12.5], [14.366026, 13.366025], [16.732052, 14.401924], [17.0, 15.901923]],
]
],
[
[
[[24.0, 24.5], [26.366024, 25.366024], [28.732052, 26.401924], [29.0, 27.901924]],
[[30.0, 30.5], [32.366028, 31.366026], [34.732048, 32.401924], [35.0, 33.901924]],
[[36.0, 36.5], [38.366024, 37.366024], [40.73205, 38.401924], [41.0, 39.901924]],
]
],
]
np.testing.assert_allclose(out, expected, atol=1e-4)
if torch.cuda.is_available():
affine = torch.as_tensor(affine, device=torch.device("cuda:0"), dtype=torch.float32)
image = torch.arange(48.0).view(2, 1, 4, 2, 3).to(device=torch.device("cuda:0"))
xform = AffineTransform(padding_mode="border", align_corners=False, mode="bilinear")
out = xform(image, affine, (3, 4, 2))
out = out.detach().cpu().numpy()
expected = [
[
[
[[0.00000006, 0.5000001], [2.3660254, 1.3660254], [4.732051, 2.4019241], [5.0, 3.9019237]],
[[6.0, 6.5], [8.366026, 7.3660254], [10.732051, 8.401924], [11.0, 9.901924]],
[[12.0, 12.5], [14.366026, 13.366025], [16.732052, 14.401924], [17.0, 15.901923]],
]
],
[
[
[[24.0, 24.5], [26.366024, 25.366024], [28.732052, 26.401924], [29.0, 27.901924]],
[[30.0, 30.5], [32.366028, 31.366026], [34.732048, 32.401924], [35.0, 33.901924]],
[[36.0, 36.5], [38.366024, 37.366024], [40.73205, 38.401924], [41.0, 39.901924]],
]
],
]
np.testing.assert_allclose(out, expected, atol=1e-4)
def test_ill_affine_transform(self):
with self.assertRaises(ValueError): # image too small
t = np.pi / 3
affine = [[1, 0, 0, 0], [0.0, np.cos(t), -np.sin(t), 0], [0, np.sin(t), np.cos(t), 0], [0, 0, 0, 1]]
affine = torch.as_tensor(affine, device=torch.device("cpu:0"), dtype=torch.float32)
xform = AffineTransform((3, 4, 2), padding_mode="border", align_corners=False, mode="bilinear")
xform(torch.as_tensor([1.0, 2.0, 3.0]), affine)
with self.assertRaises(ValueError): # output shape too small
t = np.pi / 3
affine = [[1, 0, 0, 0], [0.0, np.cos(t), -np.sin(t), 0], [0, np.sin(t), np.cos(t), 0], [0, 0, 0, 1]]
affine = torch.as_tensor(affine, device=torch.device("cpu:0"), dtype=torch.float32)
image = torch.arange(48).view(2, 1, 4, 2, 3).to(device=torch.device("cpu:0"))
xform = AffineTransform((3, 4), padding_mode="border", align_corners=False, mode="bilinear")
xform(image, affine)
with self.assertRaises(ValueError): # incorrect affine
t = np.pi / 3
affine = [[1, 0, 0, 0], [0.0, np.cos(t), -np.sin(t), 0], [0, np.sin(t), np.cos(t), 0], [0, 0, 0, 1]]
affine = torch.as_tensor(affine, device=torch.device("cpu:0"), dtype=torch.float32)
affine = affine.unsqueeze(0).unsqueeze(0)
image = torch.arange(48).view(2, 1, 4, 2, 3).to(device=torch.device("cpu:0"))
xform = AffineTransform((2, 3, 4), padding_mode="border", align_corners=False, mode="bilinear")
xform(image, affine)
with self.assertRaises(ValueError): # batch doesn't match
t = np.pi / 3
affine = [[1, 0, 0, 0], [0.0, np.cos(t), -np.sin(t), 0], [0, np.sin(t), np.cos(t), 0], [0, 0, 0, 1]]
affine = torch.as_tensor(affine, device=torch.device("cpu:0"), dtype=torch.float32)
affine = affine.unsqueeze(0)
affine = affine.repeat(3, 1, 1)
image = torch.arange(48).view(2, 1, 4, 2, 3).to(device=torch.device("cpu:0"))
xform = AffineTransform((2, 3, 4), padding_mode="border", align_corners=False, mode="bilinear")
xform(image, affine)
with self.assertRaises(RuntimeError): # input grid dtypes different
t = np.pi / 3
affine = [[1, 0, 0, 0], [0.0, np.cos(t), -np.sin(t), 0], [0, np.sin(t), np.cos(t), 0], [0, 0, 0, 1]]
affine = torch.as_tensor(affine, device=torch.device("cpu:0"), dtype=torch.float32)
affine = affine.unsqueeze(0)
affine = affine.repeat(2, 1, 1)
image = torch.arange(48).view(2, 1, 4, 2, 3).to(device=torch.device("cpu:0"), dtype=torch.int32)
xform = AffineTransform((2, 3, 4), padding_mode="border", mode="bilinear", normalized=True)
xform(image, affine)
with self.assertRaises(ValueError): # wrong affine
affine = torch.as_tensor([[1, 0, 0, 0], [0, 0, 0, 1]])
image = torch.arange(48).view(2, 1, 4, 2, 3).to(device=torch.device("cpu:0"))
xform = AffineTransform((2, 3, 4), padding_mode="border", align_corners=False, mode="bilinear")
xform(image, affine)
with self.assertRaises(RuntimeError): # dtype doesn't match
affine = torch.as_tensor([[2.0, 0.0, 0.0], [0.0, 2.0, 0.0]], dtype=torch.float64)
image = torch.arange(1.0, 13.0).view(1, 1, 3, 4).to(device=torch.device("cpu:0"))
out = AffineTransform((1, 2))(image, affine)
def test_forward_2d(self):
x = torch.rand(2, 1, 4, 4)
theta = torch.Tensor([[[0, -1, 0], [1, 0, 0]]]).repeat(2, 1, 1)
grid = torch.nn.functional.affine_grid(theta, x.size(), align_corners=False)
expected = torch.nn.functional.grid_sample(x, grid, align_corners=False)
expected = expected.detach().cpu().numpy()
actual = AffineTransform(normalized=True, reverse_indexing=False)(x, theta)
actual = actual.detach().cpu().numpy()
np.testing.assert_allclose(actual, expected)
np.testing.assert_allclose(list(theta.shape), [2, 2, 3])
theta = torch.Tensor([[0, -1, 0], [1, 0, 0]])
actual = AffineTransform(normalized=True, reverse_indexing=False)(x, theta)
actual = actual.detach().cpu().numpy()
np.testing.assert_allclose(actual, expected)
np.testing.assert_allclose(list(theta.shape), [2, 3])
theta = torch.Tensor([[[0, -1, 0], [1, 0, 0]]])
actual = AffineTransform(normalized=True, reverse_indexing=False)(x, theta)
actual = actual.detach().cpu().numpy()
np.testing.assert_allclose(actual, expected)
np.testing.assert_allclose(list(theta.shape), [1, 2, 3])
def test_forward_3d(self):
x = torch.rand(2, 1, 4, 4, 4)
theta = torch.Tensor([[[0, 0, -1, 0], [1, 0, 0, 0], [0, 0, 1, 0]]]).repeat(2, 1, 1)
grid = torch.nn.functional.affine_grid(theta, x.size(), align_corners=False)
expected = torch.nn.functional.grid_sample(x, grid, align_corners=False)
expected = expected.detach().cpu().numpy()
actual = AffineTransform(normalized=True, reverse_indexing=False)(x, theta)
actual = actual.detach().cpu().numpy()
np.testing.assert_allclose(actual, expected)
np.testing.assert_allclose(list(theta.shape), [2, 3, 4])
theta = torch.Tensor([[0, 0, -1, 0], [1, 0, 0, 0], [0, 0, 1, 0]])
actual = AffineTransform(normalized=True, reverse_indexing=False)(x, theta)
actual = actual.detach().cpu().numpy()
np.testing.assert_allclose(actual, expected)
np.testing.assert_allclose(list(theta.shape), [3, 4])
theta = torch.Tensor([[[0, 0, -1, 0], [1, 0, 0, 0], [0, 0, 1, 0]]])
actual = AffineTransform(normalized=True, reverse_indexing=False)(x, theta)
actual = actual.detach().cpu().numpy()
np.testing.assert_allclose(actual, expected)
np.testing.assert_allclose(list(theta.shape), [1, 3, 4])
if __name__ == "__main__":
unittest.main()