| import torch |
| import torch.nn as nn |
|
|
| from modules.build import HEADS_REGISTRY |
| from modules.utils import get_activation_fn |
|
|
|
|
| class BertPredictionHeadTransform(nn.Module): |
| def __init__(self, hidden_size, hidden_act='gelu'): |
| super().__init__() |
| self.dense = nn.Linear(hidden_size, hidden_size) |
| self.transform_act_fn = get_activation_fn(hidden_act) |
| self.LayerNorm = nn.LayerNorm(hidden_size) |
|
|
| def forward(self, hidden_states): |
| hidden_states = self.dense(hidden_states) |
| hidden_states = self.transform_act_fn(hidden_states) |
| hidden_states = self.LayerNorm(hidden_states) |
| return hidden_states |
|
|
|
|
| class BertLMPredictionHead(nn.Module): |
| def __init__(self, hidden_size, vocab_size): |
| super().__init__() |
| self.transform = BertPredictionHeadTransform(hidden_size=hidden_size, hidden_act='gelu') |
| self.decoder = nn.Linear(hidden_size, vocab_size, bias=False) |
| self.bias = nn.Parameter(torch.zeros(vocab_size)) |
|
|
| def forward(self, hidden_states): |
| hidden_states = self.transform(hidden_states) |
| hidden_states = self.decoder(hidden_states) + self.bias |
| return hidden_states |
|
|
|
|
| @HEADS_REGISTRY.register() |
| class PretrainHeadV1(nn.Module): |
| def __init__(self, cfg, hidden_size=768, vocab_size=30522): |
| super().__init__() |
| self.lm_pred_head = BertLMPredictionHead(hidden_size, vocab_size) |
|
|
| def forward(self, txt_embeds, **kwargs): |
| txt_lm_cls_logits = self.lm_pred_head(txt_embeds) |
| return txt_lm_cls_logits |
|
|
|
|
| @HEADS_REGISTRY.register() |
| class OVPretrainHead(nn.Module): |
| def __init__(self, cfg, hidden_size=768, vocab_size=30522, obj_vocab_size=607): |
| super().__init__() |
| self.lm_pred_head = BertLMPredictionHead(hidden_size, vocab_size) |
| self.obj_pred_head = BertLMPredictionHead(hidden_size, obj_vocab_size) |
|
|
| def forward(self, txt_embeds, obj_embeds, **kwargs): |
| txt_lm_cls_logits = self.lm_pred_head(txt_embeds) |
| obj_lm_cls_logits = self.obj_pred_head(obj_embeds) |
| return (txt_lm_cls_logits, obj_lm_cls_logits) |