| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | from typing import TYPE_CHECKING, List |
| |
|
| | from ...extras.logging import get_logger |
| |
|
| |
|
| | if TYPE_CHECKING: |
| | from transformers import PretrainedConfig, PreTrainedModel, PreTrainedTokenizer |
| |
|
| |
|
| | logger = get_logger(__name__) |
| |
|
| |
|
| | def find_all_linear_modules(model: "PreTrainedModel", freeze_vision_tower: bool) -> List[str]: |
| | r""" |
| | Finds all available modules to apply lora or galore. |
| | """ |
| | forbidden_modules = {"lm_head"} |
| |
|
| | if model.config.model_type == "chatglm": |
| | forbidden_modules.add("output_layer") |
| | elif model.config.model_type == "internlm2": |
| | forbidden_modules.add("output") |
| | elif model.config.model_type in ["llava", "paligemma"]: |
| | forbidden_modules.add("multi_modal_projector") |
| |
|
| | if freeze_vision_tower: |
| | forbidden_modules.add("vision_tower") |
| |
|
| | module_names = set() |
| | for name, module in model.named_modules(): |
| | if any(forbidden_module in name for forbidden_module in forbidden_modules): |
| | continue |
| |
|
| | if "Linear" in module.__class__.__name__ and "Embedding" not in module.__class__.__name__: |
| | module_names.add(name.split(".")[-1]) |
| |
|
| | logger.info("Found linear modules: {}".format(",".join(module_names))) |
| | return list(module_names) |
| |
|
| |
|
| | def find_expanded_modules(model: "PreTrainedModel", target_modules: List[str], num_layer_trainable: int) -> List[str]: |
| | r""" |
| | Finds the modules in the expanded blocks to apply lora. |
| | """ |
| | num_layers = getattr(model.config, "num_hidden_layers", None) |
| | if not num_layers: |
| | raise ValueError("Model was not supported.") |
| |
|
| | if num_layers % num_layer_trainable != 0: |
| | raise ValueError( |
| | "`num_layers` {} should be divisible by `num_layer_trainable` {}.".format(num_layers, num_layer_trainable) |
| | ) |
| |
|
| | stride = num_layers // num_layer_trainable |
| | trainable_layer_ids = range(stride - 1, num_layers + stride - 1, stride) |
| | trainable_layers = [".{:d}.".format(idx) for idx in trainable_layer_ids] |
| | module_names = [] |
| | for name, _ in model.named_modules(): |
| | if any(target_module in name for target_module in target_modules) and any( |
| | trainable_layer in name for trainable_layer in trainable_layers |
| | ): |
| | module_names.append(name) |
| |
|
| | logger.info("Apply lora to layers: {}".format(",".join(map(str, trainable_layer_ids)))) |
| | return module_names |
| |
|
| |
|
| | def register_autoclass(config: "PretrainedConfig", model: "PreTrainedModel", tokenizer: "PreTrainedTokenizer"): |
| | if "AutoConfig" in getattr(config, "auto_map", {}): |
| | config.__class__.register_for_auto_class() |
| | if "AutoModelForCausalLM" in getattr(config, "auto_map", {}): |
| | model.__class__.register_for_auto_class() |
| | if "AutoTokenizer" in tokenizer.init_kwargs.get("auto_map", {}): |
| | tokenizer.__class__.register_for_auto_class() |
| |
|