| import torch |
|
|
| class ReferenceOnlySimple: |
| @classmethod |
| def INPUT_TYPES(s): |
| return {"required": { "model": ("MODEL",), |
| "reference": ("LATENT",), |
| "batch_size": ("INT", {"default": 1, "min": 1, "max": 64}) |
| }} |
|
|
| RETURN_TYPES = ("MODEL", "LATENT") |
| FUNCTION = "reference_only" |
|
|
| CATEGORY = "custom_node_experiments" |
|
|
| def reference_only(self, model, reference, batch_size): |
| model_reference = model.clone() |
| size_latent = list(reference["samples"].shape) |
| size_latent[0] = batch_size |
| latent = {} |
| latent["samples"] = torch.zeros(size_latent) |
|
|
| batch = latent["samples"].shape[0] + reference["samples"].shape[0] |
| def reference_apply(q, k, v, extra_options): |
| k = k.clone().repeat(1, 2, 1) |
| offset = 0 |
| if q.shape[0] > batch: |
| offset = batch |
|
|
| for o in range(0, q.shape[0], batch): |
| for x in range(1, batch): |
| k[x + o, q.shape[1]:] = q[o,:] |
|
|
| return q, k, k |
|
|
| model_reference.set_model_attn1_patch(reference_apply) |
| out_latent = torch.cat((reference["samples"], latent["samples"])) |
| if "noise_mask" in latent: |
| mask = latent["noise_mask"] |
| else: |
| mask = torch.ones((64,64), dtype=torch.float32, device="cpu") |
|
|
| if len(mask.shape) < 3: |
| mask = mask.unsqueeze(0) |
| if mask.shape[0] < latent["samples"].shape[0]: |
| print(latent["samples"].shape, mask.shape) |
| mask = mask.repeat(latent["samples"].shape[0], 1, 1) |
|
|
| out_mask = torch.zeros((1,mask.shape[1],mask.shape[2]), dtype=torch.float32, device="cpu") |
| return (model_reference, {"samples": out_latent, "noise_mask": torch.cat((out_mask, mask))}) |
|
|
| NODE_CLASS_MAPPINGS = { |
| "ReferenceOnlySimple": ReferenceOnlySimple, |
| } |
|
|