| | """SAMPLING ONLY."""
|
| |
|
| | import torch
|
| |
|
| | from .uni_pc import NoiseScheduleVP, model_wrapper, UniPC
|
| | from modules import shared, devices
|
| |
|
| |
|
| | class UniPCSampler(object):
|
| | def __init__(self, model, **kwargs):
|
| | super().__init__()
|
| | self.model = model
|
| | to_torch = lambda x: x.clone().detach().to(torch.float32).to(model.device)
|
| | self.before_sample = None
|
| | self.after_sample = None
|
| | self.register_buffer('alphas_cumprod', to_torch(model.alphas_cumprod))
|
| |
|
| | def register_buffer(self, name, attr):
|
| | if type(attr) == torch.Tensor:
|
| | if attr.device != devices.device:
|
| | attr = attr.to(devices.device)
|
| | setattr(self, name, attr)
|
| |
|
| | def set_hooks(self, before_sample, after_sample, after_update):
|
| | self.before_sample = before_sample
|
| | self.after_sample = after_sample
|
| | self.after_update = after_update
|
| |
|
| | @torch.no_grad()
|
| | def sample(self,
|
| | S,
|
| | batch_size,
|
| | shape,
|
| | conditioning=None,
|
| | callback=None,
|
| | normals_sequence=None,
|
| | img_callback=None,
|
| | quantize_x0=False,
|
| | eta=0.,
|
| | mask=None,
|
| | x0=None,
|
| | temperature=1.,
|
| | noise_dropout=0.,
|
| | score_corrector=None,
|
| | corrector_kwargs=None,
|
| | verbose=True,
|
| | x_T=None,
|
| | log_every_t=100,
|
| | unconditional_guidance_scale=1.,
|
| | unconditional_conditioning=None,
|
| |
|
| | **kwargs
|
| | ):
|
| | if conditioning is not None:
|
| | if isinstance(conditioning, dict):
|
| | ctmp = conditioning[list(conditioning.keys())[0]]
|
| | while isinstance(ctmp, list):
|
| | ctmp = ctmp[0]
|
| | cbs = ctmp.shape[0]
|
| | if cbs != batch_size:
|
| | print(f"Warning: Got {cbs} conditionings but batch-size is {batch_size}")
|
| |
|
| | elif isinstance(conditioning, list):
|
| | for ctmp in conditioning:
|
| | if ctmp.shape[0] != batch_size:
|
| | print(f"Warning: Got {cbs} conditionings but batch-size is {batch_size}")
|
| |
|
| | else:
|
| | if conditioning.shape[0] != batch_size:
|
| | print(f"Warning: Got {conditioning.shape[0]} conditionings but batch-size is {batch_size}")
|
| |
|
| |
|
| | C, H, W = shape
|
| | size = (batch_size, C, H, W)
|
| |
|
| |
|
| | device = self.model.betas.device
|
| | if x_T is None:
|
| | img = torch.randn(size, device=device)
|
| | else:
|
| | img = x_T
|
| |
|
| | ns = NoiseScheduleVP('discrete', alphas_cumprod=self.alphas_cumprod)
|
| |
|
| |
|
| | model_type = "v" if self.model.parameterization == "v" else "noise"
|
| |
|
| | model_fn = model_wrapper(
|
| | lambda x, t, c: self.model.apply_model(x, t, c),
|
| | ns,
|
| | model_type=model_type,
|
| | guidance_type="classifier-free",
|
| |
|
| |
|
| | guidance_scale=unconditional_guidance_scale,
|
| | )
|
| |
|
| | uni_pc = UniPC(model_fn, ns, predict_x0=True, thresholding=False, variant=shared.opts.uni_pc_variant, condition=conditioning, unconditional_condition=unconditional_conditioning, before_sample=self.before_sample, after_sample=self.after_sample, after_update=self.after_update)
|
| | x = uni_pc.sample(img, steps=S, skip_type=shared.opts.uni_pc_skip_type, method="multistep", order=shared.opts.uni_pc_order, lower_order_final=shared.opts.uni_pc_lower_order_final)
|
| |
|
| | return x.to(device), None
|
| |
|