| | import os |
| | import glob |
| |
|
| | import einops |
| | import torch |
| | import torch.nn as nn |
| | import torch.nn.functional as F |
| |
|
| | from modules.build import VISION_REGISTRY |
| | from modules.layers.pointnet import PointNetPP |
| | from modules.layers.transformers import TransformerSpatialEncoderLayer |
| | from modules.utils import get_mlp_head, layer_repeat, calc_pairwise_locs, get_mixup_function |
| | from modules.weights import _init_weights_bert |
| |
|
| |
|
| | @VISION_REGISTRY.register() |
| | class PointOpenVocabEncoder(nn.Module): |
| | def __init__(self, cfg, backbone='pointnet++', hidden_size=768, path=None, freeze=False, dim_feedforward=2048, |
| | num_attention_heads=12, spatial_dim=5, num_layers=4, dim_loc=6, pairwise_rel_type='center', |
| | use_matmul_label=False, mixup_strategy=None, mixup_stage1=None, mixup_stage2=None, |
| | lang_type='bert', lang_path=None, attn_type='spatial'): |
| | super().__init__() |
| | assert backbone in ['pointnet++'] |
| |
|
| | |
| | if backbone == 'pointnet++': |
| | self.point_feature_extractor = PointNetPP( |
| | sa_n_points=[32, 16, None], |
| | sa_n_samples=[32, 32, None], |
| | sa_radii=[0.2, 0.4, None], |
| | sa_mlps=[[3, 64, 64, 128], [128, 128, 128, 256], [256, 256, 512, 768]], |
| | ) |
| | elif backbone == 'pointnext': |
| | self.point_feature_extractor = PointNextEncoder( |
| | blocks=[1, 1, 1, 1, 1, 1], |
| | strides=[1, 2, 2, 2, 2, 1], |
| | sa_layers=2, |
| | sa_use_res=True, |
| | width=32, |
| | radius=0.15, |
| | radius_scaling=1.5, |
| | mlp_head=[1024, 768] if lang_type == 'bert' else [] |
| | ) |
| |
|
| | |
| | vocab_file_name = f"scannet_607_{'bert-base-uncased' if lang_type == 'bert' else 'clip-ViT-B16'}_id.pth" |
| | self.register_buffer("text_features", torch.load(os.path.join(lang_path, vocab_file_name))) |
| | self.point_cls_head = lambda x: x @ self.text_features.t() |
| | self.dropout = nn.Dropout(0.1) |
| |
|
| | self.attn_type = attn_type |
| |
|
| | |
| | self.freeze = freeze |
| | if freeze: |
| | for p in self.parameters(): |
| | p.requires_grad = False |
| |
|
| | |
| | self.sem_cls_embed_layer = nn.Sequential(nn.Linear(hidden_size, hidden_size), |
| | nn.LayerNorm(hidden_size), |
| | nn.Dropout(0.1)) |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | self.use_matmul_label = use_matmul_label |
| | |
| | self.sem_mask_embeddings = nn.Embedding(1, 768) |
| |
|
| | |
| | if self.attn_type == 'spatial': |
| | pc_encoder_layer = TransformerSpatialEncoderLayer(hidden_size, num_attention_heads, |
| | dim_feedforward=dim_feedforward, |
| | dropout=0.1, activation='gelu', |
| | spatial_dim=spatial_dim, spatial_multihead=True, |
| | spatial_attn_fusion='cond') |
| | self.spatial_encoder = layer_repeat(pc_encoder_layer, num_layers) |
| | loc_layer = nn.Sequential( |
| | nn.Linear(dim_loc, hidden_size), |
| | nn.LayerNorm(hidden_size), |
| | ) |
| | self.loc_layers = layer_repeat(loc_layer, 1) |
| | self.pairwise_rel_type = pairwise_rel_type |
| | self.spatial_dim = spatial_dim |
| | else: |
| | pass |
| |
|
| | |
| | |
| | |
| |
|
| | |
| | self.apply(_init_weights_bert) |
| | if path is not None: |
| | |
| | |
| | |
| | |
| | print("loaded") |
| | ckpts = glob.glob(os.path.join(path, '*.bin')) |
| | if len(ckpts) != 0: |
| | for ckpt in ckpts: |
| | state_dict = torch.load(ckpt, map_location='cpu') |
| | self.load_state_dict(state_dict, strict=False) |
| | print("loaded checkpoint files") |
| | elif path.endswith('.pth'): |
| | print("loaded checkpoint file") |
| | state_dict = torch.load(path) |
| | self.load_state_dict(state_dict, strict=False) |
| | |
| | |
| | |
| | |
| |
|
| | def freeze_bn(self, m): |
| | for layer in m.modules(): |
| | if isinstance(layer, nn.BatchNorm2d): |
| | layer.eval() |
| |
|
| | def forward(self, obj_pcds, obj_locs, obj_masks, obj_sem_masks, |
| | obj_labels=None, cur_step=None, max_steps=None, **kwargs): |
| | if self.freeze: |
| | self.freeze_bn(self.point_feature_extractor) |
| |
|
| | |
| | batch_size, num_objs, _, _ = obj_pcds.size() |
| | obj_embeds = self.point_feature_extractor(einops.rearrange(obj_pcds, 'b o p d -> (b o) p d')) |
| | |
| | |
| | obj_embeds = einops.rearrange(obj_embeds, '(b o) d -> b o d', b=batch_size) |
| | obj_embeds = self.dropout(obj_embeds) |
| | if self.freeze: |
| | obj_embeds = obj_embeds.detach() |
| |
|
| | |
| | obj_sem_cls = F.softmax(self.point_cls_head(obj_embeds), dim=2).detach() |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | |
| | |
| | |
| | obj_embeds_pre = obj_embeds |
| |
|
| | |
| | if self.attn_type == 'spatial': |
| | pairwise_locs = calc_pairwise_locs(obj_locs[:, :, :3], obj_locs[:, :, 3:], |
| | pairwise_rel_type=self.pairwise_rel_type, spatial_dist_norm=True, |
| | spatial_dim=self.spatial_dim) |
| | for i, pc_layer in enumerate(self.spatial_encoder): |
| | query_pos = self.loc_layers[0](obj_locs) |
| | obj_embeds = obj_embeds + query_pos |
| | obj_embeds, self_attn_matrices = pc_layer(obj_embeds, pairwise_locs, |
| | tgt_key_padding_mask=obj_masks.logical_not()) |
| | else: |
| | pass |
| |
|
| | return obj_embeds, obj_embeds_pre, obj_sem_cls |