| | import torch.nn as nn |
| | import torch.nn.functional as F |
| | from fvcore.common.registry import Registry |
| | import torch |
| |
|
| | LOSS_REGISTRY = Registry("loss") |
| |
|
| | def og3d_loss(data_dict): |
| | return F.cross_entropy(data_dict["og3d_logits"], data_dict["tgt_object_id"].squeeze(1)) |
| |
|
| |
|
| | def og3d_multi_loss(data_dict): |
| | return F.binary_cross_entropy_with_logits( |
| | data_dict["og3d_logits"], |
| | data_dict["tgt_object_id"].float(), |
| | reduction="sum") / float(data_dict["tgt_object_id"].shape[0]) |
| |
|
| |
|
| | def txt_cls_multi_loss(data_dict): |
| | return F.binary_cross_entropy_with_logits( |
| | data_dict["txt_cls_logits"], |
| | data_dict["tgt_object_label"].float(), |
| | reduction='sum') / float(data_dict["tgt_object_label"].shape[0]) |
| |
|
| |
|
| | def obj_cls_raw_loss(data_dict): |
| | return ( |
| | F.cross_entropy( |
| | data_dict["obj_cls_raw_logits"].permute(0, 2, 1), data_dict["obj_labels"], reduction='none' |
| | ) * data_dict["obj_masks"] |
| | ).sum() / data_dict["obj_masks"].sum() |
| |
|
| |
|
| | def obj_cls_pre_loss(data_dict): |
| | return ( |
| | F.cross_entropy( |
| | data_dict["obj_cls_pre_logits"].permute(0, 2, 1), data_dict["obj_labels"], reduction='none' |
| | ) * data_dict["obj_masks"] |
| | ).sum() / data_dict["obj_masks"].sum() |
| |
|
| |
|
| | def obj_cls_post_loss(data_dict): |
| | return ( |
| | F.cross_entropy( |
| | data_dict["obj_cls_post_logits"].permute(0, 2, 1), data_dict["obj_labels"], reduction='none' |
| | ) * data_dict["obj_masks"] |
| | ).sum() / data_dict["obj_masks"].sum() |
| |
|
| |
|
| | def answer_loss(data_dict): |
| | return F.binary_cross_entropy_with_logits( |
| | data_dict["answer_scores"], data_dict["answer_label"].float(), reduction='sum' |
| | ) / data_dict["answer_scores"].shape[0] |
| |
|
| |
|
| | def lm_cls_loss(data_dict): |
| | target_labels = data_dict["masked_lm_labels"] |
| | target_labels = target_labels.view(-1, target_labels.size(-1)) if len(target_labels.size()) == 3 else target_labels |
| | return F.cross_entropy( |
| | data_dict["txt_lm_cls_logits"].permute(0, 2, 1), target_labels, ignore_index=-1 |
| | ) |
| |
|
| |
|
| | def obj_cls_pre_loss_mask(data_dict): |
| | return ( |
| | F.cross_entropy( |
| | data_dict["obj_cls_pre_logits"].permute(0, 2, 1), data_dict["obj_labels"], reduction='none' |
| | ) * data_dict["obj_masks"] * data_dict["obj_sem_masks"].logical_not() |
| | ).sum() / (data_dict["obj_masks"] * data_dict["obj_sem_masks"].logical_not()).sum() |
| |
|
| |
|
| | def obj_cls_pre_loss_unmask(data_dict): |
| | return ( |
| | F.cross_entropy( |
| | data_dict["obj_cls_pre_logits"].permute(0, 2, 1), data_dict["obj_labels"], reduction='none' |
| | ) * data_dict["obj_masks"] * data_dict["obj_sem_masks"] |
| | ).sum() / (data_dict["obj_masks"] * data_dict["obj_sem_masks"]).sum() |
| |
|
| |
|
| | def obj_cls_post_loss_mask(data_dict): |
| | return ( |
| | F.cross_entropy( |
| | data_dict["obj_cls_post_logits"].permute(0, 2, 1), data_dict["obj_labels"], reduction='none' |
| | ) * data_dict["obj_masks"] * data_dict["obj_sem_masks"].logical_not() |
| | ).sum() / (data_dict["obj_masks"] * data_dict["obj_sem_masks"].logical_not()).sum() |
| |
|
| |
|
| | def obj_cls_post_loss_unmask(data_dict): |
| | return ( |
| | F.cross_entropy( |
| | data_dict["obj_cls_post_logits"].permute(0, 2, 1), data_dict["obj_labels"], reduction='none' |
| | ) * data_dict["obj_masks"] * data_dict["obj_sem_masks"] |
| | ).sum() / (data_dict["obj_masks"] * data_dict["obj_sem_masks"]).sum() |
| |
|
| |
|
| | def obj_cls_loss(data_dict, smoothing=0.3): |
| | return ( |
| | F.cross_entropy( |
| | data_dict["obj_logits"].permute(0, 2, 1), data_dict["obj_labels"], |
| | reduction='none', label_smoothing=smoothing |
| | ) * data_dict["obj_masks"] |
| | ).sum() / data_dict["obj_masks"].sum() |
| |
|
| |
|
| | def mse_loss(data_dict): |
| | return ( |
| | ((data_dict["pred_images"] - data_dict["target_images"]) ** 2).mean() |
| | ) |
| |
|
| | class Loss(nn.Module): |
| | def __init__(self, cfg, accelerator): |
| | |
| | |
| | |
| | |
| | super().__init__() |
| | self.all_keys = list(set(cfg.model.vis_loss_list + cfg.model.loss_list)) |
| | self.selected_keys = cfg.model.loss_list |
| |
|
| | self.loss_fn = {} |
| | for k in self.all_keys: |
| | if k in globals().keys(): |
| | self.loss_fn[k] = globals()[k] |
| | print(f"Using {k} from loss.globals()") |
| | else: |
| | self.loss_fn[k] = LOSS_REGISTRY.get(k)(cfg, accelerator) |
| | setattr(self, k, self.loss_fn[k]) |
| | print(f"Using {k} from Registry {LOSS_REGISTRY._name}") |
| |
|
| | def forward(self, data_dict): |
| | all_losses = {} |
| |
|
| | |
| | if 'txt_cls_loss' in self.loss_fn and 'txt_cls_label' not in data_dict: |
| | data_dict['txt_cls_label'] = data_dict["tgt_object_label"].squeeze(1) |
| |
|
| | for k, fn in self.loss_fn.items(): |
| | |
| | cur_loss = fn(data_dict) |
| |
|
| | if isinstance(cur_loss, dict): |
| | all_losses.update(cur_loss) |
| | else: |
| | all_losses[k] = cur_loss |
| |
|
| | total_loss = sum(all_losses.values()) |
| | all_losses["total_loss"] = total_loss |
| |
|
| | return total_loss, all_losses |
| |
|
| |
|