| | """module for patching with unsloth optimizations""" |
| |
|
| | import inspect |
| | import logging |
| | import re |
| | import types |
| | from typing import Tuple |
| |
|
| | from peft import PeftModelForCausalLM |
| | from transformers.models.llama.modeling_llama import ( |
| | LlamaFlashAttention2, |
| | LlamaForCausalLM, |
| | ) |
| |
|
| | LOG = logging.getLogger("axolotl.monkeypatch.unsloth") |
| |
|
| | ORIGINAL_CEL_CODE = """ if labels is not None: |
| | # Shift so that tokens < n predict n |
| | shift_logits = logits[..., :-1, :].contiguous() |
| | shift_labels = labels[..., 1:].contiguous() |
| | # Flatten the tokens |
| | loss_fct = CrossEntropyLoss() |
| | shift_logits = shift_logits.view(-1, self.config.vocab_size) |
| | shift_labels = shift_labels.view(-1) |
| | # Enable model parallelism |
| | shift_labels = shift_labels.to(shift_logits.device) |
| | loss = loss_fct(shift_logits, shift_labels) |
| | """ |
| |
|
| | PATCHED_CEL_CODE = """ if labels is not None: |
| | shift_logits = logits[..., :-1, :].contiguous() |
| | shift_labels = labels[..., 1:].contiguous() |
| | loss = fast_cross_entropy_loss( |
| | logits = shift_logits, |
| | labels = shift_labels, |
| | ) |
| | """ |
| |
|
| | ORIGINAL_QKV_CODE = """ |
| | query_states = self.q_proj(hidden_states) |
| | key_states = self.k_proj(hidden_states) |
| | value_states = self.v_proj(hidden_states) |
| | """.lstrip( |
| | "\n" |
| | ) |
| |
|
| | PATCHED_QKV_CODE = """ |
| | query_states, key_states, value_states = self.apply_qkv(self, hidden_states) |
| | """.lstrip( |
| | "\n" |
| | ) |
| |
|
| | ORIGINAL_O_CODE = """ |
| | attn_output = self.o_proj(attn_output) |
| | """.lstrip( |
| | "\n" |
| | ) |
| |
|
| | PATCHED_O_CODE = """ |
| | attn_output = self.apply_o(self, attn_output) |
| | """.lstrip( |
| | "\n" |
| | ) |
| |
|
| |
|
| | def original_apply_qkv(self, hidden_states): |
| | query_states = self.q_proj(hidden_states) |
| | key_states = self.k_proj(hidden_states) |
| | value_states = self.v_proj(hidden_states) |
| | return query_states, key_states, value_states |
| |
|
| |
|
| | def original_apply_o(self, hidden_states): |
| | attn_output = self.o_proj(hidden_states) |
| | return attn_output |
| |
|
| |
|
| | def get_forward_code() -> str: |
| | forward = inspect.getsource(LlamaForCausalLM.forward) |
| | return forward |
| |
|
| |
|
| | def test_cel_is_patchable() -> bool: |
| | forward = get_forward_code() |
| | return ORIGINAL_CEL_CODE in forward |
| |
|
| |
|
| | def get_self_attn_code() -> str: |
| | forward = inspect.getsource(LlamaFlashAttention2.forward) |
| | return forward |
| |
|
| |
|
| | def test_self_attn_is_patchable() -> bool: |
| | qkv = get_self_attn_code() |
| | return ORIGINAL_QKV_CODE in qkv and ORIGINAL_QKV_CODE in qkv |
| |
|
| |
|
| | def integrate_cross_entropy_loss_patch(): |
| | forward = get_forward_code() |
| | LlamaForCausalLM._original_forward = forward |
| | forward, _ = detab_code(forward) |
| | assert ORIGINAL_CEL_CODE in forward, "Original forward code not found" |
| |
|
| | forward = forward.replace( |
| | "@add_start_docstrings_to_model_forward(LLAMA_INPUTS_DOCSTRING)", "" |
| | ) |
| | forward = forward.replace( |
| | "@replace_return_docstrings(output_type=CausalLMOutputWithPast, config_class=_CONFIG_FOR_DOC)", |
| | "", |
| | ) |
| | forward = forward.replace(ORIGINAL_CEL_CODE, PATCHED_CEL_CODE) |
| | forward = forward.replace( |
| | "def forward(", |
| | "def fast_cross_entropy_loss_forward(", |
| | 1, |
| | ) |
| |
|
| | |
| | import transformers.models.llama.modeling_llama |
| |
|
| | items_to_import = [] |
| | for item in dir(transformers.models.llama.modeling_llama): |
| | if item in forward: |
| | items_to_import.append(item) |
| |
|
| | exec( |
| | "from unsloth.kernels.cross_entropy_loss import fast_cross_entropy_loss", |
| | globals(), |
| | ) |
| |
|
| | exec( |
| | "from transformers.models.llama.modeling_llama import (" |
| | + ", ".join(x for x in items_to_import) |
| | + ")", |
| | globals(), |
| | ) |
| | exec(forward, globals()) |
| | print("patching unsloth fast_cross_entropy_loss") |
| | LlamaForCausalLM.forward = fast_cross_entropy_loss_forward |
| |
|
| |
|
| | def detab_code(code: str) -> Tuple[str, str]: |
| | spaces = re.match(r"([\s\t]{1,})", code).group(0) |
| | code = re.sub(r"^" + spaces, "", code, flags=re.MULTILINE) |
| | return code, spaces |
| |
|
| |
|
| | def patch_self_attn_lora(): |
| | self_attn_forward = get_self_attn_code() |
| | LlamaFlashAttention2._original_forward = ( |
| | self_attn_forward |
| | ) |
| | self_attn_forward, _ = detab_code(self_attn_forward) |
| | assert ORIGINAL_QKV_CODE in self_attn_forward, "Original qkv code not found" |
| | assert ORIGINAL_O_CODE in self_attn_forward, "Original o code not found" |
| |
|
| | self_attn_forward = self_attn_forward.replace(ORIGINAL_QKV_CODE, PATCHED_QKV_CODE) |
| | self_attn_forward = self_attn_forward.replace(ORIGINAL_O_CODE, PATCHED_O_CODE) |
| | self_attn_forward = self_attn_forward.replace( |
| | "def forward(", |
| | "def unsloth_attn_forward(", |
| | 1, |
| | ) |
| |
|
| | |
| | import transformers.models.llama.modeling_llama |
| |
|
| | items_to_import = [] |
| | for item in dir(transformers.models.llama.modeling_llama): |
| | if item in self_attn_forward: |
| | items_to_import.append(item) |
| |
|
| | exec( |
| | "from transformers.models.llama.modeling_llama import (" |
| | + ", ".join(x for x in items_to_import) |
| | + ")", |
| | globals(), |
| | ) |
| | exec(self_attn_forward, globals()) |
| | print("patching unsloth attn lora") |
| | LlamaFlashAttention2.forward = ( |
| | unsloth_attn_forward |
| | ) |
| |
|
| |
|
| | def integrate_lora_mlp_patch(peft_model: PeftModelForCausalLM): |
| | if peft_model.base_model.config.model_type in ["llama", "mistral"]: |
| | from unsloth.kernels import apply_lora_mlp_swiglu |
| |
|
| | apply_lora_mlp = apply_lora_mlp_swiglu |
| | elif peft_model.base_model.config.model_type == "gemma": |
| | from unsloth.kernels import apply_lora_mlp_geglu_approx |
| |
|
| | apply_lora_mlp = apply_lora_mlp_geglu_approx |
| | else: |
| | raise NotImplementedError( |
| | f"Model type {peft_model.base_model.config.model_type} not supported" |
| | ) |
| |
|
| | for idx, layer in enumerate(peft_model.model.model.layers): |
| | layer_modules = [ |
| | getattr(layer.mlp, linear_proj) |
| | for linear_proj in ["gate_proj", "up_proj", "down_proj"] |
| | ] |
| | is_mlp_lora = all(hasattr(module, "lora_A") for module in layer_modules) |
| | mlp_no_bias = all( |
| | getattr(module, "base_layer", module).bias is None |
| | for module in layer_modules |
| | ) |
| | mlp_not_dora = all( |
| | getattr(module, "lora_magnitude_vector", None) is None |
| | for module in layer_modules |
| | ) |
| |
|
| | if is_mlp_lora and mlp_no_bias and mlp_not_dora: |
| | layer.mlp.forward = types.MethodType(apply_lora_mlp, layer.mlp) |
| | else: |
| | logging.warning("unable to apply unsloth lora mlp patch to layer %d", idx) |
| |
|
| |
|
| | def integrate_lora_patch(peft_model: PeftModelForCausalLM, cfg): |
| | from unsloth.kernels import apply_lora_o, apply_lora_qkv |
| |
|
| | for idx, layer in enumerate(peft_model.model.model.layers): |
| | if cfg.unsloth_lora_qkv: |
| | layer_modules = [ |
| | getattr(layer.self_attn, linear_proj) |
| | for linear_proj in ["q_proj", "k_proj", "v_proj"] |
| | ] |
| | is_qkv_lora = all(hasattr(module, "lora_A") for module in layer_modules) |
| | qkv_no_bias = all( |
| | getattr(module, "base_layer", module).bias is None |
| | for module in layer_modules |
| | ) |
| | qkv_not_dora = all( |
| | getattr(module, "lora_magnitude_vector", None) is None |
| | for module in layer_modules |
| | ) |
| |
|
| | if is_qkv_lora and qkv_no_bias and qkv_not_dora: |
| | layer.self_attn.apply_qkv = apply_lora_qkv |
| | else: |
| | layer.self_attn.apply_qkv = original_apply_qkv |
| | logging.warning( |
| | "unable to apply unsloth lora qkv patch to layer %d", idx |
| | ) |
| | if cfg.unsloth_lora_o: |
| | layer_modules = [ |
| | getattr(layer.self_attn, linear_proj) for linear_proj in ["o_proj"] |
| | ] |
| | is_o_lora = all(hasattr(module, "lora_A") for module in layer_modules) |
| | o_no_bias = all( |
| | getattr(module, "base_layer", module).bias is None |
| | for module in layer_modules |
| | ) |
| | o_not_dora = all( |
| | getattr(module, "lora_magnitude_vector", None) is None |
| | for module in layer_modules |
| | ) |
| |
|
| | if is_o_lora and o_no_bias and o_not_dora: |
| | layer.self_attn.apply_o = apply_lora_o |
| | else: |
| | layer.self_attn.apply_o = original_apply_o |
| | logging.warning( |
| | "unable to apply unsloth lora o_proj patch to layer %d", idx |
| | ) |
| |
|