Stack 2.9 Training Pipeline
Complete training infrastructure to fine-tune Stack 2.9 (Qwen2.5-Coder-32B) into a super-intelligent coding model.
Overview
This pipeline provides:
- Data Preparation (
prepare_data.py) - Load, clean, format, deduplicate, and filter training data - LoRA Training (
train_lora.py) - Fine-tune with LoRA adapters - Model Merging (
merge_adapter.py) - Merge LoRA weights back to base model - AWQ Quantization (optional) - Quantize for efficient inference
Requirements
- Python 3.10+
- CUDA-compatible GPU (recommended)
- 32GB+ VRAM for base model (48GB+ recommended for training)
- 128GB+ system RAM
Quick Start
# Install dependencies
pip install -r requirements.txt
# Run complete pipeline
./run_training.sh
Individual Steps
1. Prepare Data
python prepare_data.py --config train_config.yaml
Features:
- Loads JSONL training data
- Handles multiple formats (messages, instruction/response, prompt/completion)
- Deduplication via content hash
- Quality filtering (min/max length, response presence)
- Tokenization with chat template
- 90/10 train/eval split
2. Train LoRA
python train_lora.py --config train_config.yaml
Features:
- 4-bit quantization (bitsandbytes)
- LoRA: r=64, alpha=128
- Target modules: [q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj]
- Gradient checkpointing
- Mixed precision (FP16/BF16)
- wandb/tensorboard logging
- Checkpointing and resume
3. Merge Adapter
python merge_adapter.py --config train_config.yaml
Features:
- Merges LoRA into base model
- Exports to HuggingFace format
- Optional AWQ quantization
Configuration
All hyperparameters are in train_config.yaml:
# Model
model:
name: "Qwen/Qwen2.5-Coder-32B"
torch_dtype: "bfloat16"
# Data
data:
input_path: "path/to/examples.jsonl"
max_length: 131072
# LoRA
lora:
r: 64
alpha: 128
target_modules: [...]
# Training
training:
num_epochs: 3
batch_size: 1
gradient_accumulation: 16
learning_rate: 1.0e-4
File Structure
stack-2.9-training/
βββ train_config.yaml # All hyperparameters
βββ prepare_data.py # Data preparation
βββ train_lora.py # LoRA training
βββ merge_adapter.py # Model merging
βββ run_training.sh # One-command pipeline
βββ requirements.txt # Python dependencies
βββ data/ # Processed datasets
β βββ train/
β βββ eval/
βββ output/ # Trained models
βββ stack-2.9-lora/ # LoRA adapter
βββ stack-2.9-merged/ # Merged model
βββ stack-2.9-awq/ # Quantized model
Hardware Requirements
| Config | GPU VRAM | System RAM | Training Time |
|---|---|---|---|
| Minimum | 32GB | 64GB | 12-24h |
| Recommended | 48GB+ | 128GB+ | 6-12h |
| Optimal | 80GB (A100) | 256GB+ | 4-8h |
Usage After Training
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
# Load quantized model
model = AutoModelForCausalLM.from_pretrained(
"output/stack-2.9-awq",
torch_dtype=torch.float16,
load_in_4bit=True,
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-Coder-32B")
# Generate
prompt = "Write a Python function to calculate factorial"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
output = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(output[0], skip_special_tokens=True))
Troubleshooting
Memory Issues
- Reduce
batch_sizein config - Enable gradient checkpointing
- Use 4-bit quantization
CUDA Errors
- Verify CUDA drivers:
nvidia-smi - Check PyTorch CUDA:
python -c "import torch; print(torch.cuda.is_available())"
Data Errors
- Verify JSONL format
- Check required columns: messages OR instruction/response
License
Provided as-is for educational and research purposes.