| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| import unittest |
| from typing import Tuple |
|
|
| import numpy as np |
| import torch |
| from parameterized import parameterized |
|
|
| from monai.metrics import compute_confusion_metric |
|
|
| metric_name = "true positive rate" |
| |
| TEST_CASES_2_CLS_CLF = [ |
| [ |
| { |
| "y_pred": torch.tensor([[0.4], [0.4], [0.4], [0.6], [0.9], [0.9]]), |
| "y": torch.tensor([[1], [0], [0], [1], [1], [1]]), |
| "bin_mode": "threshold", |
| "metric_name": metric_name, |
| }, |
| 0.75, |
| ], |
| [ |
| { |
| "y_pred": torch.tensor([[0], [0], [0], [1], [1], [1]]), |
| "y": torch.tensor([[1], [0], [0], [1], [1], [1]]), |
| "bin_mode": None, |
| "metric_name": metric_name, |
| }, |
| 0.75, |
| ], |
| ] |
|
|
| |
| average_list = ["micro", "macro", "weighted"] |
| multi_class_result_list = [0.4, 0.33333334, 0.4] |
| TEST_CASES_M_CLS_CLF = [] |
| y_pred = torch.tensor([[0.4, 0.8, 1], [0.4, 0.8, 0], [0.4, 0.2, 0.7], [0.6, 0.1, 0.2], [0.2, 0.1, 0.8]]) |
| y = torch.tensor([[1], [0], [2], [1], [2]]) |
| for i in range(len(average_list)): |
| average = average_list[i] |
| test_case = [ |
| { |
| "y_pred": y_pred, |
| "y": y, |
| "to_onehot_y": True, |
| "metric_name": metric_name, |
| "bin_mode": "mutually_exclusive", |
| "average": average, |
| }, |
| multi_class_result_list[i], |
| ] |
| TEST_CASES_M_CLS_CLF.append(test_case) |
|
|
|
|
| |
| multi_label_result_list = [0.75, 0.80555556, 0.75] |
| TEST_CASES_M_LABEL_CLF = [] |
| y_pred = torch.tensor([[0.4, 0.8, 1], [0.4, 0.8, 0], [0.4, 0.2, 0.7], [0.6, 0.1, 0.2], [0.2, 0.1, 0.8]]) |
| y = torch.tensor([[0, 1, 1], [0, 1, 0], [0, 1, 1], [1, 0, 1], [0, 0, 1]]) |
| for i in range(len(average_list)): |
| average = average_list[i] |
| test_case = [ |
| { |
| "y_pred": y_pred, |
| "y": y, |
| "metric_name": metric_name, |
| "bin_mode": "threshold", |
| "bin_threshold": [0.3, 0.6, 0.5], |
| "average": average, |
| }, |
| multi_label_result_list[i], |
| ] |
| TEST_CASES_M_LABEL_CLF.append(test_case) |
|
|
|
|
| |
| def produce_seg_input(shape: Tuple): |
| y = torch.cat([torch.ones(shape), torch.zeros(shape)]) |
| y_pred = torch.rand_like(y) * 0.5 |
| return y, y_pred |
|
|
|
|
| seg_result_list_2d = [0, 0, 0, 1, 1, 1] |
| basic_shape: Tuple |
| basic_shape = (5, 5, 3, 3) |
| y, y_org = produce_seg_input(basic_shape) |
| TEST_CASES_2D_SEG = [] |
| ct = 0 |
| for y_pred in [y_org, 1 - y_org]: |
| for i in range(len(average_list)): |
| average = average_list[i] |
| test_case = [ |
| {"y_pred": y_pred, "y": y, "metric_name": metric_name, "bin_mode": "threshold", "average": average}, |
| seg_result_list_2d[ct], |
| ] |
| ct += 1 |
| TEST_CASES_2D_SEG.append(test_case) |
|
|
|
|
| |
| seg_result_list_3d = [1, 1, 1, 0, 0, 0] |
| basic_shape = (2, 5, 3, 3, 3) |
| y, y_org = produce_seg_input(basic_shape) |
| TEST_CASES_3D_SEG = [] |
| ct = 0 |
| for y_pred in [y_org, -y_org]: |
| for i in range(len(average_list)): |
| average = average_list[i] |
| test_case = [ |
| { |
| "y_pred": y_pred, |
| "y": y, |
| "activation": "sigmoid", |
| "metric_name": metric_name, |
| "bin_mode": "threshold", |
| "average": average, |
| }, |
| seg_result_list_3d[ct], |
| ] |
| ct += 1 |
| TEST_CASES_3D_SEG.append(test_case) |
|
|
|
|
| class TestComputeTprClf(unittest.TestCase): |
| @parameterized.expand(TEST_CASES_2_CLS_CLF + TEST_CASES_M_CLS_CLF + TEST_CASES_M_LABEL_CLF) |
| def test_value(self, input_data, expected_value): |
| result = compute_confusion_metric(**input_data) |
| np.testing.assert_allclose(expected_value, result, rtol=1e-7) |
|
|
|
|
| class TestComputeTprSeg(unittest.TestCase): |
| @parameterized.expand(TEST_CASES_2D_SEG + TEST_CASES_3D_SEG) |
| def test_value(self, input_data, expected_value): |
| result = compute_confusion_metric(**input_data) |
| self.assertEqual(expected_value, result) |
|
|
|
|
| if __name__ == "__main__": |
| unittest.main() |
|
|