| import torch |
|
|
|
|
| def convert_gt_to_one_hot(gt_segments, gt_labels, num_classes): |
| """convert the gt from class index to one hot encoding. this is for multi class case.""" |
|
|
| gt_segments_unique, gt_labels_onehot = [], [] |
| for gt_segment, gt_label in zip(gt_segments, gt_labels): |
| if len(gt_segment) > 0: |
| bbox_unique, inverse_indices = torch.unique(gt_segment, dim=0, return_inverse=True) |
| label_unique = [] |
| for i in range(bbox_unique.shape[0]): |
| label = torch.nn.functional.one_hot( |
| gt_label[inverse_indices == i].long(), |
| num_classes=num_classes, |
| ) |
| label_unique.append(label.sum(dim=0).to(gt_label.device)) |
| label_unique = torch.stack(label_unique) |
| else: |
| bbox_unique, label_unique = [], [] |
| gt_segments_unique.append(bbox_unique) |
| gt_labels_onehot.append(label_unique) |
| |
| |
| return gt_segments_unique, gt_labels_onehot |
|
|