| """ |
| Export the trained model to HuggingFace-compatible format. |
| |
| Creates: |
| - model.safetensors (weights) |
| - config.json (architecture config) |
| - generation_config.json |
| - tokenizer.json, tokenizer_config.json, special_tokens_map.json |
| """ |
|
|
| import os |
| import sys |
| import json |
| import torch |
| from collections import OrderedDict |
| from safetensors.torch import save_file |
|
|
| sys.path.insert(0, os.path.dirname(os.path.abspath(__file__))) |
| from model.config import ModelConfig |
| from model.transformer import Transformer |
| from model.data import get_tokenizer |
|
|
| CHECKPOINT = "/jfs/deepak-kumar/checkpoints_dpo/dpo_final.pt" |
| OUTPUT_DIR = "/home/jovyan/training/hf_model" |
|
|
| os.makedirs(OUTPUT_DIR, exist_ok=True) |
|
|
| print("=" * 60) |
| print(" EXPORTING MODEL TO HUGGING FACE FORMAT") |
| print("=" * 60) |
|
|
| |
| print("\n[1/4] Loading checkpoint...") |
| tokenizer = get_tokenizer() |
| special_tokens = ["<|user|>", "<|assistant|>", "<|end|>"] |
| vocab = tokenizer.get_vocab() |
| new_tokens = [t for t in special_tokens if t not in vocab] |
| if new_tokens: |
| tokenizer.add_tokens(new_tokens, special_tokens=True) |
|
|
| model_config = ModelConfig() |
| model_config.vocab_size = len(tokenizer) |
|
|
| model = Transformer(model_config) |
| ckpt = torch.load(CHECKPOINT, map_location="cpu", weights_only=False) |
| model.load_state_dict(ckpt["model"]) |
| step = ckpt.get("step", 0) |
| del ckpt |
| print(f" Loaded DPO model (step {step}, vocab {model_config.vocab_size})") |
|
|
| |
| print("\n[2/4] Converting weights to safetensors...") |
|
|
| state_dict = model.state_dict() |
| hf_state = OrderedDict() |
|
|
| KEY_MAP = { |
| "tok_embeddings.weight": "model.embed_tokens.weight", |
| "norm.weight": "model.norm.weight", |
| "output.weight": "lm_head.weight", |
| } |
|
|
| for key, tensor in state_dict.items(): |
| if key in KEY_MAP: |
| hf_state[KEY_MAP[key]] = tensor |
| continue |
|
|
| if key.startswith("layers."): |
| parts = key.split(".") |
| layer_idx = parts[1] |
| rest = ".".join(parts[2:]) |
|
|
| layer_map = { |
| "attention_norm.weight": f"model.layers.{layer_idx}.input_layernorm.weight", |
| "ffn_norm.weight": f"model.layers.{layer_idx}.post_attention_layernorm.weight", |
| "attention.wq.weight": f"model.layers.{layer_idx}.self_attn.q_proj.weight", |
| "attention.wk.weight": f"model.layers.{layer_idx}.self_attn.k_proj.weight", |
| "attention.wv.weight": f"model.layers.{layer_idx}.self_attn.v_proj.weight", |
| "attention.wo.weight": f"model.layers.{layer_idx}.self_attn.o_proj.weight", |
| "ffn.w_gate.weight": f"model.layers.{layer_idx}.mlp.gate_proj.weight", |
| "ffn.w_up.weight": f"model.layers.{layer_idx}.mlp.up_proj.weight", |
| "ffn.w_down.weight": f"model.layers.{layer_idx}.mlp.down_proj.weight", |
| } |
|
|
| if rest in layer_map: |
| hf_state[layer_map[rest]] = tensor |
| else: |
| print(f" WARNING: unmapped key {key}") |
| hf_state[key] = tensor |
| elif key == "freqs_cis": |
| continue |
| else: |
| print(f" WARNING: unmapped key {key}") |
| hf_state[key] = tensor |
|
|
| |
| for k in hf_state: |
| if hf_state[k].dtype == torch.float32: |
| hf_state[k] = hf_state[k].to(torch.bfloat16) |
|
|
| safetensors_path = os.path.join(OUTPUT_DIR, "model.safetensors") |
| save_file(hf_state, safetensors_path) |
| size_gb = os.path.getsize(safetensors_path) / 1e9 |
| print(f" Saved {len(hf_state)} tensors -> {safetensors_path} ({size_gb:.2f} GB)") |
|
|
| |
| print("\n[3/4] Writing config files...") |
|
|
| config_json = { |
| "architectures": ["LlamaForCausalLM"], |
| "model_type": "llama", |
| "vocab_size": model_config.vocab_size, |
| "hidden_size": model_config.hidden_dim, |
| "intermediate_size": model_config.intermediate_dim, |
| "num_hidden_layers": model_config.num_layers, |
| "num_attention_heads": model_config.num_attention_heads, |
| "num_key_value_heads": model_config.num_kv_heads, |
| "max_position_embeddings": model_config.max_seq_len, |
| "rope_theta": model_config.rope_theta, |
| "rms_norm_eps": model_config.rms_norm_eps, |
| "hidden_act": "silu", |
| "initializer_range": 0.02, |
| "tie_word_embeddings": False, |
| "torch_dtype": "bfloat16", |
| "transformers_version": "4.40.0", |
| "use_cache": True, |
| "bos_token_id": tokenizer.bos_token_id, |
| "eos_token_id": tokenizer.eos_token_id, |
| "pad_token_id": tokenizer.pad_token_id, |
| } |
|
|
| with open(os.path.join(OUTPUT_DIR, "config.json"), "w") as f: |
| json.dump(config_json, f, indent=2) |
| print(" config.json") |
|
|
| gen_config = { |
| "bos_token_id": tokenizer.bos_token_id, |
| "eos_token_id": tokenizer.eos_token_id, |
| "pad_token_id": tokenizer.pad_token_id, |
| "do_sample": True, |
| "temperature": 0.7, |
| "top_k": 50, |
| "top_p": 0.9, |
| "repetition_penalty": 1.15, |
| "max_new_tokens": 512, |
| "transformers_version": "4.40.0", |
| } |
|
|
| with open(os.path.join(OUTPUT_DIR, "generation_config.json"), "w") as f: |
| json.dump(gen_config, f, indent=2) |
| print(" generation_config.json") |
|
|
| |
| print("\n[4/4] Exporting tokenizer...") |
| tokenizer.save_pretrained(OUTPUT_DIR) |
| print(" Tokenizer files saved") |
|
|
| print("\n" + "=" * 60) |
| print(" EXPORT COMPLETE -> " + OUTPUT_DIR) |
| print("=" * 60) |
| print("\nFiles:") |
| for f in sorted(os.listdir(OUTPUT_DIR)): |
| size = os.path.getsize(os.path.join(OUTPUT_DIR, f)) |
| if size > 1e6: |
| print(f" {f:40s} {size/1e6:.1f} MB") |
| else: |
| print(f" {f:40s} {size/1e3:.1f} KB") |
|
|