| | import torch |
| | import torch.nn.functional as F |
| | from torch import nn |
| |
|
| | from modules.build import HEADS_REGISTRY |
| |
|
| |
|
| | class FC(nn.Module): |
| | def __init__(self, in_size, out_size, pdrop=0., use_gelu=True): |
| | super(FC, self).__init__() |
| | self.pdrop = pdrop |
| | self.use_gelu = use_gelu |
| | self.linear = nn.Linear(in_size, out_size) |
| | if use_gelu: |
| | |
| | self.gelu = nn.GELU() |
| | if pdrop > 0: |
| | self.dropout = nn.Dropout(pdrop) |
| |
|
| | def forward(self, x): |
| | x = self.linear(x) |
| | if self.use_gelu: |
| | |
| | x = self.gelu(x) |
| | if self.pdrop > 0: |
| | x = self.dropout(x) |
| | return x |
| |
|
| |
|
| | class MLP(nn.Module): |
| | def __init__(self, in_size, mid_size, out_size, pdrop=0., use_gelu=True): |
| | super().__init__() |
| | self.fc = FC(in_size, mid_size, pdrop=pdrop, use_gelu=use_gelu) |
| | self.linear = nn.Linear(mid_size, out_size) |
| |
|
| | def forward(self, x): |
| | return self.linear(self.fc(x)) |
| |
|
| |
|
| | class AttFlat(nn.Module): |
| | def __init__(self, hidden_size, flat_mlp_size=512, flat_glimpses=1, flat_out_size=1024, pdrop=0.1): |
| | super().__init__() |
| | self.mlp = MLP( |
| | in_size=hidden_size, |
| | mid_size=flat_mlp_size, |
| | out_size=flat_glimpses, |
| | pdrop=pdrop, |
| | use_gelu=True |
| | ) |
| | self.flat_glimpses = flat_glimpses |
| | self.linear_merge = nn.Linear( |
| | hidden_size * flat_glimpses, |
| | flat_out_size |
| | ) |
| |
|
| | def forward(self, x, x_mask): |
| | att = self.mlp(x) |
| | if x_mask is not None: |
| | |
| | att = att.masked_fill(x_mask.unsqueeze(2), -1e9) |
| | att = F.softmax(att, dim=1) |
| | att_list = [] |
| | for i in range(self.flat_glimpses): |
| | att_list.append( |
| | torch.sum(att[:, :, i: i + 1] * x, dim=1) |
| | ) |
| | x_atted = torch.cat(att_list, dim=1) |
| | x_atted = self.linear_merge(x_atted) |
| | return x_atted |
| |
|
| |
|
| | @HEADS_REGISTRY.register() |
| | class QAHeadV1(nn.Module): |
| | def __init__(self, cfg, hidden_size=768, mlp_size=256, glimpse=1, flat_out_size=512, num_answers=8864): |
| | super().__init__() |
| | image_embed_dim = 512 |
| | text_embed_dim = 1024 |
| | |
| | self.attflat_pm = AttFlat(image_embed_dim, mlp_size, glimpse, flat_out_size, 0.1) |
| | |
| | self.attflat_lang = AttFlat(text_embed_dim, mlp_size, glimpse, flat_out_size, 0.1) |
| | self.answer_cls = nn.Sequential( |
| | nn.Linear(flat_out_size, hidden_size), |
| | nn.GELU(), |
| | nn.Dropout(0.3), |
| | nn.Linear(hidden_size, num_answers) |
| | ) |
| | self.fusion_norm = nn.LayerNorm(flat_out_size) |
| |
|
| | def forward(self, pm_embeds, txt_embeds, txt_masks): |
| | |
| | pm_feat = self.attflat_pm(pm_embeds, None) |
| | |
| | lang_feat = self.attflat_lang(txt_embeds, txt_masks) |
| |
|
| | |
| | fuse_feat = self.fusion_norm(lang_feat + pm_feat) |
| | answer_scores = self.answer_cls(fuse_feat) |
| | return answer_scores |