walidsobhie-code
refactor: Squeeze folders further - cleaner structure
65888d5

Stack 2.9 Training Pipeline

Complete training infrastructure to fine-tune Stack 2.9 (Qwen2.5-Coder-32B) into a super-intelligent coding model.

Overview

This pipeline provides:

  1. Data Preparation (prepare_data.py) - Load, clean, format, deduplicate, and filter training data
  2. LoRA Training (train_lora.py) - Fine-tune with LoRA adapters
  3. Model Merging (merge_adapter.py) - Merge LoRA weights back to base model
  4. AWQ Quantization (optional) - Quantize for efficient inference

Requirements

  • Python 3.10+
  • CUDA-compatible GPU (recommended)
  • 32GB+ VRAM for base model (48GB+ recommended for training)
  • 128GB+ system RAM

Quick Start

# Install dependencies
pip install -r requirements.txt

# Run complete pipeline
./run_training.sh

Individual Steps

1. Prepare Data

python prepare_data.py --config train_config.yaml

Features:

  • Loads JSONL training data
  • Handles multiple formats (messages, instruction/response, prompt/completion)
  • Deduplication via content hash
  • Quality filtering (min/max length, response presence)
  • Tokenization with chat template
  • 90/10 train/eval split

2. Train LoRA

python train_lora.py --config train_config.yaml

Features:

  • 4-bit quantization (bitsandbytes)
  • LoRA: r=64, alpha=128
  • Target modules: [q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj]
  • Gradient checkpointing
  • Mixed precision (FP16/BF16)
  • wandb/tensorboard logging
  • Checkpointing and resume

3. Merge Adapter

python merge_adapter.py --config train_config.yaml

Features:

  • Merges LoRA into base model
  • Exports to HuggingFace format
  • Optional AWQ quantization

Configuration

All hyperparameters are in train_config.yaml:

# Model
model:
  name: "Qwen/Qwen2.5-Coder-32B"
  torch_dtype: "bfloat16"

# Data
data:
  input_path: "path/to/examples.jsonl"
  max_length: 131072

# LoRA
lora:
  r: 64
  alpha: 128
  target_modules: [...]

# Training
training:
  num_epochs: 3
  batch_size: 1
  gradient_accumulation: 16
  learning_rate: 1.0e-4

File Structure

stack-2.9-training/
β”œβ”€β”€ train_config.yaml      # All hyperparameters
β”œβ”€β”€ prepare_data.py      # Data preparation
β”œβ”€β”€ train_lora.py        # LoRA training
β”œβ”€β”€ merge_adapter.py    # Model merging
β”œβ”€β”€ run_training.sh      # One-command pipeline
β”œβ”€β”€ requirements.txt    # Python dependencies
β”œβ”€β”€ data/                # Processed datasets
β”‚   β”œβ”€β”€ train/
β”‚   └── eval/
└── output/              # Trained models
    β”œβ”€β”€ stack-2.9-lora/    # LoRA adapter
    β”œβ”€β”€ stack-2.9-merged/  # Merged model
    └── stack-2.9-awq/     # Quantized model

Hardware Requirements

Config GPU VRAM System RAM Training Time
Minimum 32GB 64GB 12-24h
Recommended 48GB+ 128GB+ 6-12h
Optimal 80GB (A100) 256GB+ 4-8h

Usage After Training

from transformers import AutoModelForCausalLM, AutoTokenizer
import torch

# Load quantized model
model = AutoModelForCausalLM.from_pretrained(
    "output/stack-2.9-awq",
    torch_dtype=torch.float16,
    load_in_4bit=True,
    device_map="auto"
)

tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-Coder-32B")

# Generate
prompt = "Write a Python function to calculate factorial"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
output = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(output[0], skip_special_tokens=True))

Troubleshooting

Memory Issues

  • Reduce batch_size in config
  • Enable gradient checkpointing
  • Use 4-bit quantization

CUDA Errors

  • Verify CUDA drivers: nvidia-smi
  • Check PyTorch CUDA: python -c "import torch; print(torch.cuda.is_available())"

Data Errors

  • Verify JSONL format
  • Check required columns: messages OR instruction/response

License

Provided as-is for educational and research purposes.