xcdata / code /policy_models /module /diffusion_decoder.py
fjwwjf151's picture
Upload folder using huggingface_hub
b506011 verified
import einops
from policy_models.module.transformers.transformer_blocks import *
class SinusoidalPosEmb(nn.Module):
def __init__(self, dim):
super().__init__()
self.dim = dim
def forward(self, x):
device = x.device
half_dim = self.dim // 2
emb = math.log(10000) / (half_dim - 1)
emb = torch.exp(torch.arange(half_dim, device=device) * -emb)
emb = x[:, None] * emb[None, :]
emb = torch.cat((emb.sin(), emb.cos()), dim=-1)
return emb
logger = logging.getLogger(__name__)
def return_model_parameters_in_millions(model):
num_params = sum(p.numel() for p in model.parameters() if p.requires_grad)
num_params_in_millions = round(num_params / 1_000_000, 2)
return num_params_in_millions
class DiffusionTransformer(nn.Module):
"""the full GPT score model, with a context size of block_size"""
def __init__(
self,
obs_dim: int,
goal_dim: int,
device: str,
n_obs_token: int,
goal_conditioned: bool,
action_dim: int,
proprio_dim: int,
embed_dim: int,
embed_pdrob: float,
attn_pdrop: float,
resid_pdrop: float,
mlp_pdrop: float,
n_dec_layers: int,
n_enc_layers: int,
n_heads: int,
goal_seq_len: int,
obs_seq_len: int,
action_seq_len: int,
goal_drop: float = 0.1,
bias=False,
use_mlp_goal: bool = False,
use_rot_embed: bool = False,
rotary_xpos: bool = False,
linear_output: bool = True,
use_noise_encoder: bool = False,
use_ada_conditioning: bool = True,
):
super().__init__()
self.device = device
self.goal_conditioned = goal_conditioned
self.obs_dim = obs_dim
self.embed_dim = embed_dim
self.n_obs_token = n_obs_token
self.use_ada_conditioning = use_ada_conditioning
if self.goal_conditioned:
block_size = goal_seq_len + action_seq_len + obs_seq_len * self.n_obs_token + 2
else:
block_size = action_seq_len + obs_seq_len * self.n_obs_token + 2
self.action_seq_len = action_seq_len
if self.goal_conditioned:
seq_size = goal_seq_len + obs_seq_len * self.n_obs_token + action_seq_len
else:
seq_size = obs_seq_len * self.n_obs_token + action_seq_len
print(f"obs dim: {obs_dim}, goal_dim: {goal_dim}, action_dim: {action_dim}, proprio_dim: {proprio_dim}")
self.tok_emb = nn.Linear(obs_dim, embed_dim)
if use_mlp_goal:
self.goal_emb = nn.Sequential(
nn.Linear(goal_dim, embed_dim * 2),
nn.GELU(),
nn.Linear(embed_dim * 2, embed_dim)
)
else:
self.goal_emb = nn.Linear(goal_dim, embed_dim)
if use_mlp_goal:
self.lang_emb = nn.Sequential(
nn.Linear(goal_dim, embed_dim * 2),
nn.GELU(),
nn.Linear(embed_dim * 2, embed_dim)
)
else:
self.lang_emb = nn.Linear(goal_dim, embed_dim)
if not self.goal_conditioned:
for param in self.lang_emb.parameters():
param.requires_grad = False
for param in self.goal_emb.parameters():
param.requires_grad = False
self.pos_emb = nn.Parameter(torch.zeros(1, seq_size, embed_dim))
print('seq_size:',seq_size)
self.drop = nn.Dropout(embed_pdrob)
self.proprio_drop = nn.Dropout(0.5)
self.cond_mask_prob = goal_drop
self.use_rot_embed = use_rot_embed
self.action_dim = action_dim
self.obs_dim = obs_dim
self.embed_dim = embed_dim
self.latent_encoder_emb = None
self.encoder = TransformerEncoder(
embed_dim=embed_dim,
n_heads=n_heads,
attn_pdrop=attn_pdrop,
resid_pdrop=resid_pdrop,
n_layers=n_enc_layers,
block_size=block_size,
bias=bias,
use_rot_embed=use_rot_embed,
rotary_xpos=rotary_xpos,
mlp_pdrop=mlp_pdrop,
)
self.decoder = TransformerFiLMDecoder(
embed_dim=embed_dim,
n_heads=n_heads,
attn_pdrop=attn_pdrop,
resid_pdrop=resid_pdrop,
n_layers=n_dec_layers,
film_cond_dim=embed_dim,
block_size=block_size,
bias=bias,
use_rot_embed=use_rot_embed,
rotary_xpos=rotary_xpos,
mlp_pdrop=mlp_pdrop,
use_cross_attention=True,
use_noise_encoder=use_noise_encoder,
)
self.latent_encoder_emb = None
self.proprio_emb = nn.Sequential(
nn.Linear(proprio_dim, embed_dim * 2),
nn.Mish(),
nn.Linear(embed_dim * 2, embed_dim),
).to(self.device)
self.block_size = block_size
self.goal_seq_len = goal_seq_len
self.obs_seq_len = obs_seq_len
self.sigma_emb = nn.Sequential(
SinusoidalPosEmb(embed_dim),
nn.Linear(embed_dim, embed_dim * 2),
nn.Mish(),
nn.Linear(embed_dim * 2, embed_dim),
).to(self.device)
self.action_emb = nn.Linear(action_dim, embed_dim)
if linear_output:
self.action_pred = nn.Linear(embed_dim, self.action_dim)
else:
self.action_pred = nn.Sequential(
nn.Linear(embed_dim, 100),
nn.GELU(),
nn.Linear(100, self.action_dim)
)
self.apply(self._init_weights)
logger.info(f'Number of encoder parameters: {return_model_parameters_in_millions(self.encoder)}')
logger.info(f'Number of decoder parameters: {return_model_parameters_in_millions(self.decoder)}')
logger.info(
"number of parameters: %e", sum(p.numel() for p in self.parameters())
)
def get_block_size(self):
return self.block_size
def _init_weights(self, module):
if isinstance(module, (nn.Linear, nn.Embedding)):
torch.nn.init.normal_(module.weight, mean=0.0, std=0.02)
if isinstance(module, nn.Linear) and module.bias is not None:
torch.nn.init.zeros_(module.bias)
elif isinstance(module, nn.LayerNorm):
torch.nn.init.zeros_(module.bias)
torch.nn.init.ones_(module.weight)
elif isinstance(module, DiffusionTransformer):
torch.nn.init.normal_(module.pos_emb, mean=0.0, std=0.02)
def forward(self, states, actions, goals, sigma, uncond: Optional[bool] = False):
context = self.forward_enc_only(states, actions, goals, sigma, uncond)
pred_actions = self.forward_dec_only(context, actions, sigma)
return pred_actions
def forward_enc_only(self, states, actions=None, goals=None, sigma=None, uncond: Optional[bool] = False):
emb_t = self.process_sigma_embeddings(sigma) if not self.use_ada_conditioning else None
goals = self.preprocess_goals(goals, states['state_images'].size(1), uncond)
state_embed, proprio_embed = self.process_state_embeddings(states)
goal_embed = self.process_goal_embeddings(goals)
input_seq = self.concatenate_inputs(emb_t, goal_embed, state_embed, proprio_embed, uncond)
context = self.encoder(input_seq)
self.latent_encoder_emb = context
return context
def forward_dec_only(self, context, actions, sigma):
emb_t = self.process_sigma_embeddings(sigma)
action_embed = self.action_emb(actions)
action_x = self.drop(action_embed)
x = self.decoder(action_x, emb_t, context)
pred_actions = self.action_pred(x)
return pred_actions
def process_sigma_embeddings(self, sigma):
sigmas = sigma.log() / 4
sigmas = einops.rearrange(sigmas, 'b -> b 1')
emb_t = self.sigma_emb(sigmas)
if len(emb_t.shape) == 2:
emb_t = einops.rearrange(emb_t, 'b d -> b 1 d')
return emb_t
def preprocess_goals(self, goals, states_length, uncond=False):
if len(goals.shape) == 2:
goals = einops.rearrange(goals, 'b d -> b 1 d')
if goals.shape[1] == states_length and self.goal_seq_len == 1:
goals = goals[:, 0, :]
goals = einops.rearrange(goals, 'b d -> b 1 d')
if goals.shape[-1] == 2 * self.obs_dim:
goals = goals[:, :, :self.obs_dim]
if self.training:
goals = self.mask_cond(goals)
if uncond:
goals = torch.zeros_like(goals).to(self.device)
return goals
def process_state_embeddings(self, states):
states_global = self.tok_emb(states['state_images'])
if 'state_obs' in states:
proprio_embed = self.proprio_emb(states['state_obs'])
else:
proprio_embed = None
return states_global, proprio_embed
def process_goal_embeddings(self, goals):
goal_embed = self.lang_emb(goals)
return goal_embed
def apply_position_embeddings(self, goal_embed, state_embed, action_embed, proprio_embed, t):
pos_len = t + self.goal_seq_len + self.action_seq_len - 1
position_embeddings = self.pos_emb[:, :pos_len, :]
goal_x = self.drop(goal_embed + position_embeddings[:, :self.goal_seq_len, :])
state_x = self.drop(state_embed + position_embeddings[:, self.goal_seq_len:(self.goal_seq_len + t), :])
action_x = self.drop(action_embed + position_embeddings[:, (self.goal_seq_len + t - 1):, :])
proprio_x = self.drop(proprio_embed + position_embeddings[:, self.goal_seq_len:(self.goal_seq_len + t), :]) if proprio_embed is not None else None
return goal_x, state_x, action_x, proprio_x
def concatenate_inputs(self, emb_t, goal_x, state_x, proprio_x, uncond=False):
input_seq_components = [state_x]
if self.goal_conditioned:
input_seq_components.insert(0, goal_x)
if proprio_x is not None:
input_seq_components.append(proprio_x)
#else:
# if not self.goal_conditioned:
# input_seq_components.append(self.drop(goal_x))
input_seq = torch.cat(input_seq_components, dim=1)
return input_seq
def mask_cond(self, cond, force_mask=False):
bs, t, d = cond.shape
if force_mask:
return torch.zeros_like(cond)
elif self.training and self.cond_mask_prob > 0.:
mask = torch.bernoulli(torch.ones((bs, t, d), device=cond.device) * self.cond_mask_prob)
return cond * (1. - mask)
else:
return cond
def get_params(self):
return self.parameters()