| | import torch |
| |
|
| | from contextlib import contextmanager |
| | from typing import Union, Tuple |
| |
|
| |
|
| | _size_2_t = Union[int, Tuple[int, int]] |
| |
|
| |
|
| | class LinearWithLoRA(torch.nn.Module): |
| | def __init__( |
| | self, |
| | in_features: int, |
| | out_features: int, |
| | bias: bool = True, |
| | device=None, |
| | dtype=None) -> None: |
| | super().__init__() |
| | self.weight_module = None |
| | self.up = None |
| | self.down = None |
| | self.bias = None |
| | self.in_features = in_features |
| | self.out_features = out_features |
| | self.device = device |
| | self.dtype = dtype |
| | self.weight = None |
| |
|
| | def bind_lora(self, weight_module): |
| | self.weight_module = [weight_module] |
| |
|
| | def unbind_lora(self): |
| | if self.up is not None and self.down is not None: |
| | self.weight_module = None |
| |
|
| | def get_original_weight(self): |
| | if self.weight_module is None: |
| | return None |
| | return self.weight_module[0].weight |
| |
|
| | def forward(self, x): |
| | if self.weight is not None: |
| | return torch.nn.functional.linear(x, self.weight.to(x), |
| | self.bias.to(x) if self.bias is not None else None) |
| |
|
| | original_weight = self.get_original_weight() |
| |
|
| | if original_weight is None: |
| | return None |
| |
|
| | if self.up is not None and self.down is not None: |
| | weight = original_weight.to(x) + torch.mm(self.up, self.down).to(x) |
| | else: |
| | weight = original_weight.to(x) |
| |
|
| | return torch.nn.functional.linear(x, weight, self.bias.to(x) if self.bias is not None else None) |
| |
|
| |
|
| | class Conv2dWithLoRA(torch.nn.Module): |
| | def __init__( |
| | self, |
| | in_channels: int, |
| | out_channels: int, |
| | kernel_size: _size_2_t, |
| | stride: _size_2_t = 1, |
| | padding: Union[str, _size_2_t] = 0, |
| | dilation: _size_2_t = 1, |
| | groups: int = 1, |
| | bias: bool = True, |
| | padding_mode: str = 'zeros', |
| | device=None, |
| | dtype=None |
| | ) -> None: |
| | super().__init__() |
| | self.stride = stride |
| | self.padding = padding |
| | self.dilation = dilation |
| | self.groups = groups |
| | self.weight_module = None |
| | self.bias = None |
| | self.up = None |
| | self.down = None |
| | self.in_channels = in_channels |
| | self.out_channels = out_channels |
| | self.kernel_size = kernel_size |
| | self.padding_mode = padding_mode |
| | self.device = device |
| | self.dtype = dtype |
| | self.weight = None |
| |
|
| | def bind_lora(self, weight_module): |
| | self.weight_module = [weight_module] |
| |
|
| | def unbind_lora(self): |
| | if self.up is not None and self.down is not None: |
| | self.weight_module = None |
| |
|
| | def get_original_weight(self): |
| | if self.weight_module is None: |
| | return None |
| | return self.weight_module[0].weight |
| |
|
| | def forward(self, x): |
| | if self.weight is not None: |
| | return torch.nn.functional.conv2d(x, self.weight.to(x), self.bias.to(x) if self.bias is not None else None, |
| | self.stride, self.padding, self.dilation, self.groups) |
| |
|
| | original_weight = self.get_original_weight() |
| |
|
| | if original_weight is None: |
| | return None |
| |
|
| | if self.up is not None and self.down is not None: |
| | weight = original_weight.to(x) + torch.mm(self.up.flatten(start_dim=1), self.down.flatten(start_dim=1)).reshape(original_weight.shape).to(x) |
| | else: |
| | weight = original_weight.to(x) |
| |
|
| | return torch.nn.functional.conv2d(x, weight, self.bias.to(x) if self.bias is not None else None, |
| | self.stride, self.padding, self.dilation, self.groups) |
| |
|
| |
|
| | @contextmanager |
| | def controlnet_lora_hijack(): |
| | linear, conv2d = torch.nn.Linear, torch.nn.Conv2d |
| | torch.nn.Linear, torch.nn.Conv2d = LinearWithLoRA, Conv2dWithLoRA |
| | try: |
| | yield |
| | finally: |
| | torch.nn.Linear, torch.nn.Conv2d = linear, conv2d |
| |
|
| |
|
| | def recursive_set(obj, key, value): |
| | if obj is None: |
| | return |
| | if '.' in key: |
| | k1, k2 = key.split('.', 1) |
| | recursive_set(getattr(obj, k1, None), k2, value) |
| | else: |
| | setattr(obj, key, value) |
| |
|
| |
|
| | def force_load_state_dict(model, state_dict): |
| | for k in list(state_dict.keys()): |
| | recursive_set(model, k, torch.nn.Parameter(state_dict[k])) |
| | del state_dict[k] |
| | return |
| |
|
| |
|
| | def recursive_bind_lora(obj, key, value): |
| | if obj is None: |
| | return |
| | if '.' in key: |
| | k1, k2 = key.split('.', 1) |
| | recursive_bind_lora(getattr(obj, k1, None), k2, value) |
| | else: |
| | target = getattr(obj, key, None) |
| | if target is not None and hasattr(target, 'bind_lora'): |
| | target.bind_lora(value) |
| |
|
| |
|
| | def recursive_get(obj, key): |
| | if obj is None: |
| | return |
| | if '.' in key: |
| | k1, k2 = key.split('.', 1) |
| | return recursive_get(getattr(obj, k1, None), k2) |
| | else: |
| | return getattr(obj, key, None) |
| |
|
| |
|
| | def bind_control_lora(base_model, control_lora_model): |
| | sd = base_model.state_dict() |
| | keys = list(sd.keys()) |
| | keys = list(set([k.rsplit('.', 1)[0] for k in keys])) |
| | module_dict = {k: recursive_get(base_model, k) for k in keys} |
| | for k, v in module_dict.items(): |
| | recursive_bind_lora(control_lora_model, k, v) |
| |
|
| |
|
| | def torch_dfs(model: torch.nn.Module): |
| | result = [model] |
| | for child in model.children(): |
| | result += torch_dfs(child) |
| | return result |
| |
|
| |
|
| | def unbind_control_lora(control_lora_model): |
| | for m in torch_dfs(control_lora_model): |
| | if hasattr(m, 'unbind_lora'): |
| | m.unbind_lora() |
| | return |
| |
|