| | import torch |
| | import torch.nn as nn |
| | import json |
| | from pathlib import Path |
| |
|
| | import clip |
| | from transformers import BertConfig, BertModel, BertTokenizer |
| | from einops import rearrange |
| |
|
| | from model.build import MODEL_REGISTRY, BaseModel |
| | from modules.layers.pointnet import PointNetPP |
| | from modules.utils import get_mlp_head |
| | from optim.utils import no_decay_param_group |
| |
|
| |
|
| | @MODEL_REGISTRY.register() |
| | class ObjCls(BaseModel): |
| | def __init__(self, cfg): |
| | super().__init__(cfg) |
| | self.cfg = cfg |
| | self.model_name = cfg.model.get("model_name", "pointnext") |
| | self.language_type = cfg.model.get("language_type", "clip") |
| | self.pre_extract_path = cfg.model.get("pre_extract_path", None) |
| |
|
| | cls_in_channel = 512 if self.language_type == "clip" else 768 |
| | self.point_feature_extractor = PointNetPP( |
| | sa_n_points=[32, 16, None], |
| | sa_n_samples=[32, 32, None], |
| | sa_radii=[0.2, 0.4, None], |
| | sa_mlps=[[3, 64, 64, 128], [128, 128, 128, 256], [256, 256, 512, cls_in_channel]], |
| | ) |
| |
|
| | if cfg.num_gpu > 1: |
| | self.point_feature_extractor = torch.nn.SyncBatchNorm.convert_sync_batchnorm(self.point_feature_extractor) |
| |
|
| | if not cfg.model.open_vocab: |
| | cls_hidden = cfg.model.get("cls_hidden", 1024) |
| | num_classes = cfg.model.num_classes |
| | self.cls_head = get_mlp_head(cls_in_channel, cls_hidden, num_classes) |
| | else: |
| | if self.pre_extract_path is not None: |
| | file_name = f"scannet_607_{'clip-ViT-B16' if self.language_type == 'clip' else 'bert-base-uncased'}_id.pth" |
| | self.register_buffer("text_embeds", torch.load(Path(self.pre_extract_path) / file_name).float()) |
| | else: |
| | self.int2cat = json.load(open(cfg.model.vocab_path, "r")) |
| | if self.language_type == "clip": |
| | self.clip_head = clip.load("ViT-B/16") |
| | self.text_embeds = self.clip_head.encode_text(clip.tokenize(self.int2cat)).detach() |
| | elif self.language_type == "bert": |
| | self.tokenizer = BertTokenizer.from_pretrained("bert-base-uncased", do_lower_case=True) |
| | self.bert_config = BertConfig( |
| | hidden_size=768, num_hidden_layers=3, num_attention=12, type_vocab_size=2 |
| | ) |
| | self.model = BertModel.from_pretrained("bert-base-uncased", config=self.bert_config) |
| | self.encoded_input = self.tokenizer( |
| | self.int2cat, padding=True, truncation=True, add_special_tokens=True, return_tensors="pt" |
| | ) |
| | self.text_embeds = self.model(**self.encoded_input).last_hidden_state |
| | self.text_embeds = self.text_embeds.detach() |
| | else: |
| | raise NotImplementedError |
| | self.dropout = nn.Dropout(0.1) |
| |
|
| | def forward(self, data_dict): |
| | |
| | if 'cur_step' not in data_dict.keys(): |
| | data_dict['cur_step'] = 1 |
| | data_dict['total_steps'] = 1 |
| |
|
| | obj_pcds = data_dict["obj_fts"] |
| | batch_size, num_objs, _, _ = obj_pcds.size() |
| | if self.model_name == "pointnext": |
| | obj_locs = rearrange(obj_pcds[..., :3], 'b o p d -> (b o) p d') |
| | obj_fts = rearrange(obj_pcds[..., 3:], 'b o p d -> (b o) d p').contiguous() |
| | obj_embeds = self.point_feature_extractor(obj_locs, obj_fts, type="cls") |
| | elif self.model_name == "pointnet++": |
| | obj_pcds = rearrange(obj_pcds, 'b o p d -> (b o) p d') |
| | obj_embeds = self.point_feature_extractor(obj_pcds) |
| | elif self.model_name == "pointmlp": |
| | obj_pcds = rearrange(obj_pcds, 'b o p d -> (b o) p d') |
| | obj_embeds = self.point_feature_extractor(obj_pcds) |
| | obj_embeds = self.dropout(obj_embeds) |
| | if self.cfg.model.open_vocab: |
| | logits = obj_embeds @ self.text_embeds.t() |
| | data_dict["obj_logits"] = rearrange(logits, '(b o) c -> b o c', b=batch_size) |
| | else: |
| | data_dict["obj_logits"] = rearrange(self.cls_head(obj_embeds), '(b o) d -> b o d', b=batch_size) |
| | return data_dict |
| |
|
| | def get_opt_params(self): |
| | optimizer_grouped_parameters = [] |
| | optimizer_grouped_parameters.append({ |
| | "params": self.parameters(), |
| | "weight_decay": self.cfg.solver.get("weight_decay", 0.0), |
| | "lr": self.cfg.solver.lr |
| | }) |
| | return optimizer_grouped_parameters |
| |
|